General Language Understanding Evaluation (GLUE) benchmark is a collection of nine natural language understanding tasks, including single-sentence tasks CoLA and SST-2, similarity and paraphrasing tasks MRPC, STS-B and QQP, and natural language inference tasks MNLI, QNLI, RTE and WNLI.
Source: Align, Mask and Select: A Simple Method for Incorporating Commonsense Knowledge into Language Representation ModelsPaper | Code | Results | Date | Stars |
---|