OpenCodePapers

linguistic-acceptability-on-cola

Linguistic Acceptability
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAccuracyMCCModelNameReleaseDate
Acceptability Judgements via Examining the Topology of Attention Maps✓ Link88.6%En-BERT + TDA + PCA2022-05-19
Can BERT eat RuCoLA? Topological Data Analysis to Explain✓ Link88.2%0.726BERT+TDA2023-04-04
Can BERT eat RuCoLA? Topological Data Analysis to Explain✓ Link87.3%0.695RoBERTa+TDA2023-04-04
tasksource: A Dataset Harmonization Framework for Streamlined NLP Multi-Task Learning and Evaluation✓ Link87.15%deberta-v3-base+tasksource2023-01-14
Entailment as Few-Shot Learner✓ Link86.4%RoBERTa-large 355M + Entailment as Few-shot Learner2021-04-29
Not all layers are equally as important: Every Layer Counts BERT82.7LTG-BERT-base 98M2023-11-03
Not all layers are equally as important: Every Layer Counts BERT82.6ELC-BERT-base 98M2023-11-03
Acceptability Judgements via Examining the Topology of Attention Maps✓ Link82.1%0.565En-BERT + TDA2022-05-19
FNet: Mixing Tokens with Fourier Transforms✓ Link78%FNet-Large2021-05-09
Not all layers are equally as important: Every Layer Counts BERT77.6LTG-BERT-small 24M2023-11-03
Not all layers are equally as important: Every Layer Counts BERT76.1ELC-BERT-small 24M2023-11-03
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link70.8%T5-11B2019-10-23
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding69.2%StructBERTRoBERTa ensemble2019-08-13
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations✓ Link69.1%ALBERT2019-09-26
XLNet: Generalized Autoregressive Pretraining for Language Understanding✓ Link69%XLNet (single model)2019-06-19
Learning to Encode Position for Transformer with Continuous Dynamical Model✓ Link69%FLOATER-large2020-03-13
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale✓ Link68.6%RoBERTa-large 355M (MLP quantized vector-wise, fine-tuned)2022-08-15
Multi-Task Deep Neural Networks for Natural Language Understanding✓ Link68.4%MT-DNN2019-01-31
[]()68.2%ELECTRA
RoBERTa: A Robustly Optimized BERT Pretraining Approach✓ Link67.8%RoBERTa (ensemble)2019-07-26
A Statistical Framework for Low-bitwidth Training of Deep Neural Networks✓ Link67.5PSQ (Chen et al., 2020)2020-10-27
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link67.1%T5-XL 3B2019-10-23
Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT65.1Q-BERT (Shen et al., 2020)2019-09-12
Q8BERT: Quantized 8Bit BERT✓ Link65.0Q8BERT (Zafrir et al., 2019)2019-10-14
SpanBERT: Improving Pre-training by Representing and Predicting Spans✓ Link64.3%SpanBERT2019-07-24
CLEAR: Contrastive Learning for Sentence Representation64.3%MLM+ del-span+ reorder2020-12-31
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding✓ Link63.5%ERNIE 2.0 Large2019-07-29
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link61.2%T5-Large 770M2019-10-23
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding✓ Link60.5%BERT-LARGE2018-10-11
data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language✓ Link60.3%data2vec2022-02-07
RealFormer: Transformer Likes Residual Attention✓ Link59.83%RealFormer2020-12-21
Big Bird: Transformers for Longer Sequences✓ Link58.5%BigBird2020-07-28
How to Train BERT with an Academic Budget✓ Link57.124hBERT2021-04-15
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding✓ Link55.2%ERNIE 2.0 Base2019-07-29
ERNIE: Enhanced Language Representation with Informative Entities✓ Link52.3%ERNIE2019-05-17
Charformer: Fast Character Transformers via Gradient-based Subword Tokenization✓ Link51.8%Charformer-Tall2021-06-23
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link51.1%T5-Base2019-10-23
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter✓ Link49.1%DistilBERT 66M2019-10-02
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?✓ Link46.5%SqueezeBERT2020-06-19
TinyBERT: Distilling BERT for Natural Language Understanding✓ Link43.3%TinyBERT-4 14.5M2019-09-23
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link41.0%T5-Small2019-10-23
LM-CPPF: Paraphrasing-Guided Data Augmentation for Contrastive Prompt-Based Few-Shot Fine-Tuning✓ Link14.1%LM-CPPF RoBERTa-base2023-05-29
RuCoLA: Russian Corpus of Linguistic Acceptability✓ Link0.6RemBERT2022-10-23