OpenCodePapers

natural-language-inference-on-wnli

Natural Language Inference
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAccuracyModelNameReleaseDate
[]()95.9Turing NLR v5 XXL 5.4B (fine-tuned)
DeBERTa: Decoding-enhanced BERT with Disentangled Attention✓ Link94.5DeBERTa2020-06-05
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link93.2T5-XXL 11B2019-10-23
XLNet: Generalized Autoregressive Pretraining for Language Understanding✓ Link92.5XLNet2019-06-19
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations✓ Link91.8ALBERT2019-09-26
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link89.7T5-XL 3B2019-10-23
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding89.7StructBERTRoBERTa ensemble2019-08-13
A Hybrid Neural Network Model for Commonsense Reasoning✓ Link89HNNensemble2019-07-27
RoBERTa: A Robustly Optimized BERT Pretraining Approach✓ Link89RoBERTa (ensemble)2019-07-26
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link85.6T5-Large 770M2019-10-23
A Hybrid Neural Network Model for Commonsense Reasoning✓ Link83.6HNN2019-07-27
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link78.8T5-Base 220M2019-10-23
A Surprisingly Robust Trick for Winograd Schema Challenge✓ Link74.7BERTwiki 340M (fine-tuned on WSCR)2019-05-15
Finetuned Language Models Are Zero-Shot Learners✓ Link74.6FLAN 137B (zero-shot)2021-09-03
A Surprisingly Robust Trick for Winograd Schema Challenge✓ Link71.9BERT-large 340M (fine-tuned on WSCR)2019-05-15
A Surprisingly Robust Trick for Winograd Schema Challenge✓ Link70.5BERT-base 110M (fine-tuned on WSCR)2019-05-15
Finetuned Language Models Are Zero-Shot Learners✓ Link70.4FLAN 137B (few-shot, k=4)2021-09-03
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link69.2T5-Small 60M2019-10-23
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding✓ Link67.8ERNIE 2.0 Large2019-07-29
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?✓ Link65.1SqueezeBERT2020-06-19
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding✓ Link65.1BERT-large 340M2018-10-11
RWKV: Reinventing RNNs for the Transformer Era✓ Link49.3RWKV-4-Raven-14B2023-05-22
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter✓ Link44.4DistilBERT 66M2019-10-02