OpenCodePapers

natural-language-inference-on-qnli

Natural Language Inference
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAccuracyModelNameReleaseDate
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations✓ Link99.2%ALBERT2019-09-26
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding99.2%StructBERTRoBERTa ensemble2019-08-13
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link99.2%ALICE2019-11-08
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link99.2%MT-DNN-SMART2019-11-08
RoBERTa: A Robustly Optimized BERT Pretraining Approach✓ Link98.9%RoBERTa (ensemble)2019-07-26
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link96.7%T5-11B2019-10-23
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link96.3%T5-3B2019-10-23
DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing✓ Link96%DeBERTaV3large2021-11-18
[]()95.4%ELECTRA
DeBERTa: Decoding-enhanced BERT with Disentangled Attention✓ Link95.3%DeBERTa (large)2020-06-05
XLNet: Generalized Autoregressive Pretraining for Language Understanding✓ Link94.9%XLNet (single model)2019-06-19
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link94.8%T5-Large 770M2019-10-23
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale✓ Link94.7%RoBERTa-large 355M (MLP quantized vector-wise, fine-tuned)2022-08-15
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding✓ Link94.6%ERNIE 2.0 Large2019-07-29
A Statistical Framework for Low-bitwidth Training of Deep Neural Networks✓ Link94.5PSQ (Chen et al., 2020)2020-10-27
Entailment as Few-Shot Learner✓ Link94.5%RoBERTa-large 355M + Entailment as Few-shot Learner2021-04-29
SpanBERT: Improving Pre-training by Representing and Predicting Spans✓ Link94.3%SpanBERT2019-07-24
TRANS-BLSTM: Transformer with Bidirectional LSTM for Language Understanding94.08%TRANS-BLSTM2020-03-16
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link93.7%T5-Base2019-10-23
Adversarial Self-Attention for Language Understanding✓ Link93.6%ASA + RoBERTa2022-06-25
CLEAR: Contrastive Learning for Sentence Representation93.4%MLM+ subs+ del-span2020-12-31
Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT93.0Q-BERT (Shen et al., 2020)2019-09-12
Q8BERT: Quantized 8Bit BERT✓ Link93.0Q8BERT (Zafrir et al., 2019)2019-10-14
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding✓ Link92.9%ERNIE 2.0 Base2019-07-29
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding✓ Link92.7%BERT-LARGE2018-10-11
Big Bird: Transformers for Longer Sequences✓ Link92.2%BigBird2020-07-28
RealFormer: Transformer Likes Residual Attention✓ Link91.89%RealFormer2020-12-21
Adversarial Self-Attention for Language Understanding✓ Link91.4%ASA + BERT-base2022-06-25
ERNIE: Enhanced Language Representation with Informative Entities✓ Link91.3%ERNIE2019-05-17
data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language✓ Link91.1%data2vec2022-02-07
Charformer: Fast Character Transformers via Gradient-based Subword Tokenization✓ Link91.0%Charformer-Tall2021-06-23
How to Train BERT with an Academic Budget✓ Link90.624hBERT2021-04-15
SenseBERT: Driving Some Sense into BERT90.6%SenseBERT-base 110M2019-08-15
TinyBERT: Distilling BERT for Natural Language Understanding✓ Link90.4%TinyBERT-6 67M2019-09-23
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link90.3%T5-Small2019-10-23
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter✓ Link90.2%DistilBERT 66M2019-10-02
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?✓ Link90.1%SqueezeBERT2020-06-19
Nyströmformer: A Nyström-Based Algorithm for Approximating Self-Attention✓ Link88.7%Nyströmformer2021-02-07
TinyBERT: Distilling BERT for Natural Language Understanding✓ Link87.7%TinyBERT-4 14.5M2019-09-23
FNet: Mixing Tokens with Fourier Transforms✓ Link85%FNet-Large2021-05-09
LM-CPPF: Paraphrasing-Guided Data Augmentation for Contrastive Prompt-Based Few-Shot Fine-Tuning✓ Link70.2%LM-CPPF RoBERTa-base2023-05-29
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ LinkSMART-BERT2019-11-08
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ LinkSMARTRoBERTa2019-11-08