OpenCodePapers

sentiment-analysis-on-sst-2-binary

Sentiment Analysis
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAccuracyDev AccuracyAttack Success RateModelNameReleaseDate
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link97.5T5-11B2019-10-23
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link97.5MT-DNN-SMART2019-11-08
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link97.4T5-3B2019-10-23
Muppet: Massive Multi-task Representations with Pre-Finetuning✓ Link97.4MUPPET Roberta Large2021-01-26
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations✓ Link97.1ALBERT2019-09-26
StructBERT: Incorporating Language Structures into Pre-training for Deep Language Understanding97.1StructBERTRoBERTa ensemble2019-08-13
XLNet: Generalized Autoregressive Pretraining for Language Understanding✓ Link97XLNet (single model)2019-06-19
ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators✓ Link96.9ELECTRA2020-03-23
Entailment as Few-Shot Learner✓ Link96.9RoBERTa-large 355M + Entailment as Few-shot Learner2021-04-29
XLNet: Generalized Autoregressive Pretraining for Language Understanding✓ Link96.8XLNet-Large (ensemble)2019-06-19
Learning to Encode Position for Transformer with Continuous Dynamical Model✓ Link96.7FLOATER-large2020-03-13
Muppet: Massive Multi-task Representations with Pre-Finetuning✓ Link96.7MUPPET Roberta base2021-01-26
RoBERTa: A Robustly Optimized BERT Pretraining Approach✓ Link96.7RoBERTa (ensemble)2019-07-26
DeBERTa: Decoding-enhanced BERT with Disentangled Attention✓ Link96.5DeBERTa (large)2020-06-05
Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding✓ Link96.5MT-DNN-ensemble2019-04-20
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale✓ Link96.4RoBERTa-large 355M (MLP quantized vector-wise, fine-tuned)2022-08-15
Adversarial Self-Attention for Language Understanding✓ Link96.3ASA + RoBERTa2022-06-25
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link96.3T5-Large 770M2019-10-23
Training Complex Models with Multi-Task Weak Supervision✓ Link96.2Snorkel MeTaL(ensemble)2018-10-05
A Statistical Framework for Low-bitwidth Training of Deep Neural Networks✓ Link96.2PSQ (Chen et al., 2020)2020-10-27
An Algorithm for Routing Vectors in Sequences✓ Link96.0Heinsen Routing + RoBERTa-large2022-11-20
Multi-Task Deep Neural Networks for Natural Language Understanding✓ Link95.6MT-DNN2019-01-31
An Algorithm for Routing Capsules in All Domains✓ Link95.6Heinsen Routing + GPT-22019-11-02
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link95.2T5-Base2019-10-23
ERNIE 2.0: A Continual Pre-training Framework for Language Understanding✓ Link95ERNIE 2.0 Base2019-07-29
Dual Contrastive Learning: Text Classification via Label-Aware Data Augmentation✓ Link94.91RoBERTa+DualCL2022-01-21
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding✓ Link94.9BERT-LARGE2018-10-11
SubRegWeigh: Effective and Efficient Annotation Weighing with Subword Regularization✓ Link94.84RoBERTa + SubRegWeigh (K-means)2024-09-10
SpanBERT: Improving Pre-training by Representing and Predicting Spans✓ Link94.8SpanBERT2019-07-24
Pay Attention to MLPs✓ Link94.8gMLP-large2021-05-17
Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT94.8Q-BERT (Shen et al., 2020)2019-09-12
Q8BERT: Quantized 8Bit BERT✓ Link94.7Q8BERT (Zafrir et al., 2019)2019-10-14
Cloze-driven Pretraining of Self-attention Networks94.6CNN Large2019-03-19
Big Bird: Transformers for Longer Sequences✓ Link94.6BigBird2020-07-28
CLEAR: Contrastive Learning for Sentence Representation94.5MLM+ del-word+ reorder2020-12-31
Adversarial Self-Attention for Language Understanding✓ Link94.1ASA + BERT-base2022-06-25
RealFormer: Transformer Likes Residual Attention✓ Link94.04RealFormer2020-12-21
FNet: Mixing Tokens with Fourier Transforms✓ Link94FNet-Large2021-05-09
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link93.6MT-DNN2019-11-08
ERNIE: Enhanced Language Representation with Informative Entities✓ Link93.5ERNIE2019-05-17
GPU Kernels for Block-Sparse Weights✓ Link93.2Block-sparse LSTM2017-12-01
LM-CPPF: Paraphrasing-Guided Data Augmentation for Contrastive Prompt-Based Few-Shot Fine-Tuning✓ Link93.2LM-CPPF RoBERTa-base2023-05-29
TinyBERT: Distilling BERT for Natural Language Understanding✓ Link93.1TinyBERT-6 67M2019-09-23
How to Train BERT with an Academic Budget✓ Link93.024hBERT2021-04-15
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link93SMART+BERT-BASE2019-11-08
TinyBERT: Distilling BERT for Natural Language Understanding✓ Link92.6TinyBERT-4 14.5M2019-09-23
Learning to Generate Reviews and Discovering Sentiment✓ Link91.8bmLSTM2017-04-05
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link91.8T5-Small2019-10-23
A La Carte Embedding: Cheap but Effective Induction of Semantic Feature Vectors✓ Link91.7byte mLSTM72018-05-14
Pay Attention when Required✓ Link91.6PAR BERT Base2020-09-09
Charformer: Fast Character Transformers via Gradient-based Subword Tokenization✓ Link91.6Charformer-Base2021-06-23
SqueezeBERT: What can computer vision teach NLP about efficient neural networks?✓ Link91.4SqueezeBERT2020-06-19
Nyströmformer: A Nyström-Based Algorithm for Approximating Self-Attention✓ Link91.4Nyströmformer2021-02-07
Cell-aware Stacked LSTMs for Modeling Sentences91.3Bi-CAS-LSTM2018-09-07
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter✓ Link91.3DistilBERT 66M2019-10-02
On the Role of Text Preprocessing in Neural Network Architectures: An Evaluation Study on Text Categorization and Sentiment Analysis✓ Link91.2CNN2017-07-06
Improved Sentence Modeling using Suffix Bidirectional LSTM91.2Suffix BiLSTM2018-05-18
Fine-grained Sentiment Classification using BERT✓ Link91.2BERT Base2019-10-04
Practical Text Classification With Large Pre-Trained Language Models✓ Link90.9Transformer (finetune)2018-12-04
Distilling Task-Specific Knowledge from BERT into Simple Neural Networks✓ Link90.7Single layer bilstm distilled from BERT2019-03-28
Learned in Translation: Contextualized Word Vectors✓ Link90.3BCN+Char+CoVe2017-08-01
Convolutional Neural Networks with Recurrent Neural Filters✓ Link90.0CNN-RNF-LSTM2018-08-28
Neural Semantic Encoders✓ Link89.7Neural Semantic Encoder2016-07-14
Text Classification Improved by Integrating Bidirectional LSTM with Two-dimensional Max Pooling✓ Link89.5BLSTM-2DCNN2016-11-21
Harnessing Deep Neural Networks with Logic Rules✓ Link89.3CNN + Logic rules2016-03-21
Ask Me Anything: Dynamic Memory Networks for Natural Language Processing✓ Link88.6DMN [ankit16]2015-06-24
Convolutional Neural Networks for Sentence Classification✓ Link88.1CNN-multichannel [kim2013]2014-08-25
Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks✓ Link88.0Consistency Tree LSTM with tuned Glove vectors [tai2015improved]2015-02-28
A C-LSTM Neural Network for Text Classification✓ Link87.8C-LSTM2015-11-27
Message Passing Attention Networks for Document Understanding✓ Link87.75MPAD-path2019-08-17
Information Aggregation via Dynamic Routing for Sequence Encoding✓ Link87.6Standard DR-AGG2018-06-05
Universal Sentence Encoder✓ Link87.21USE_T+CNN (lrn w.e.) 2018-03-29
Information Aggregation via Dynamic Routing for Sequence Encoding✓ Link87.2Reverse DR-AGG2018-06-05
A Helping Hand: Transfer Learning for Deep Sentiment Analysis86.99DC-MCNN2018-07-01
The Pupil Has Become the Master: Teacher-Student Model-Based Word Embedding Distillation with Ensemble Learning✓ Link86.95STM+TSED+PT+2L2019-05-31
Investigating Capsule Networks with Dynamic Routing for Text Classification✓ Link86.8Capsule-B 2018-03-29
Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks✓ Link86.32-layer LSTM [tai2015improved]2015-02-28
Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms✓ Link84.3SWEM-concat2018-05-24
Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank✓ Link82.9MV-RNN2013-10-01
Emo2Vec: Learning Generalized Emotion Representation by Multi-task Training✓ Link82.3GloVe+Emo2Vec2018-09-12
Emo2Vec: Learning Generalized Emotion Representation by Multi-task Training✓ Link81.2Emo2Vec2018-09-12
Task-oriented Word Embedding for Text Classification✓ Link78.8ToWE-CBOW2018-08-01
Exploring Joint Neural Model for Sentence Level Discourse Parsing and Sentiment Analysis 54.72 Joined Model Multi-tasking2017-08-01
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link96.9SMARTRoBERTa2019-11-08
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link96.1SMART-MT-DNN2019-11-08
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization✓ Link93.0SMART-BERT2019-11-08
Fine-mixing: Mitigating Backdoors in Fine-tuned Language Models✓ Link100Word+ES (Scratch)2022-10-18