OpenCodePapers

question-answering-on-squad11-dev

Question Answering
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeEMF1ModelNameReleaseDate
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link90.0695.64T5-11B2019-10-23
LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention✓ Link89.8LUKE2020-10-02
Dice Loss for Data-imbalanced NLP Tasks✓ Link89.7995.77XLNet+DSC2019-11-07
XLNet: Generalized Autoregressive Pretraining for Language Understanding✓ Link89.795.1XLNet (single model)2019-06-19
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link88.5394.95T5-3B2019-10-23
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link86.6693.79T5-Large 770M2019-10-23
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding✓ Link86.292.2BERT-LARGE (Ensemble+TriviaQA)2018-10-11
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link85.4492.08T5-Base2019-10-23
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding✓ Link84.291.1BERT-LARGE (Single+TriviaQA)2018-10-11
Prune Once for All: Sparse Pre-Trained Language Models✓ Link83.3590.2BERT-Large-uncased-PruneOFA (90% unstruct sparse)2021-11-10
Prune Once for All: Sparse Pre-Trained Language Models✓ Link83.2290.02BERT-Large-uncased-PruneOFA (90% unstruct sparse, QAT Int8)2021-11-10
Prune Once for All: Sparse Pre-Trained Language Models✓ Link81.188.42BERT-Base-uncased-PruneOFA (85% unstruct sparse)2021-11-10
Prune Once for All: Sparse Pre-Trained Language Models✓ Link80.8488.24BERT-Base-uncased-PruneOFA (85% unstruct sparse, QAT Int8)2021-11-10
Prune Once for All: Sparse Pre-Trained Language Models✓ Link79.8387.25BERT-Base-uncased-PruneOFA (90% unstruct sparse)2021-11-10
TinyBERT: Distilling BERT for Natural Language Understanding✓ Link79.787.5TinyBERT-6 67M2019-09-23
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer✓ Link79.187.24T5-Small2019-10-23
Reinforced Mnemonic Reader for Machine Reading Comprehension✓ Link78.9 86.3R.M-Reader (single)2017-05-08
Learning Dense Representations of Phrases at Scale✓ Link78.386.3DensePhrases2020-12-23
Prune Once for All: Sparse Pre-Trained Language Models✓ Link78.185.82DistilBERT-uncased-PruneOFA (85% unstruct sparse)2021-11-10
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter✓ Link77.7DistilBERT2019-10-02
Prune Once for All: Sparse Pre-Trained Language Models✓ Link77.0385.13DistilBERT-uncased-PruneOFA (85% unstruct sparse, QAT Int8)2021-11-10
Prune Once for All: Sparse Pre-Trained Language Models✓ Link76.9184.82DistilBERT-uncased-PruneOFA (90% unstruct sparse)2021-11-10
Explicit Utilization of General Knowledge in Machine Reading Comprehension76.784.9 KAR2018-09-10
Stochastic Answer Networks for Machine Reading Comprehension✓ Link76.23584.056SAN (single)2017-12-10
Prune Once for All: Sparse Pre-Trained Language Models✓ Link75.6283.87DistilBERT-uncased-PruneOFA (90% unstruct sparse, QAT Int8)2021-11-10
FusionNet: Fusing via Fully-Aware Attention with Application to Machine Comprehension✓ Link75.383.6FusionNet2017-11-16
QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension✓ Link75.183.8QANet (data aug x3)2018-04-23
QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension✓ Link74.583.2QANet (data aug x2)2018-04-23
DCN+: Mixed Objective and Deep Residual Coattention for Question Answering✓ Link74.583.1DCN+ (single)2017-10-31
QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension✓ Link73.682.7QANet2018-04-23
Phase Conductor on Multi-layered Attentions for Machine Comprehension72.181.4PhaseCond (single)2017-10-28
Simple Recurrent Units for Highly Parallelizable Recurrence✓ Link71.480.2SRU2017-09-08
Smarnet: Teaching Machines to Read and Comprehend Like Human71.36280.183Smarnet2017-10-08
Learned in Translation: Contextualized Word Vectors✓ Link71.379.9DCN (Char + CoVe)2017-08-01
Gated Self-Matching Networks for Reading Comprehension and Question Answering71.179.5R-NET (single)2017-07-01
Ruminating Reader: Reasoning with Gated Multi-Hop Attention70.679.5Ruminating Reader2017-04-24
Making Neural QA as Simple as Possible but not Simpler✓ Link70.378.5FastQAExt (beam-size 5)2017-03-14
Reading Wikipedia to Answer Open-Domain Questions✓ Link69.578.8DrQA (Document Reader only)2017-03-31
Exploring Question Understanding and Adaptation in Neural-Network-Based Question Answering69.1078.38jNet (TreeLSTM adaptation, QTLa, K=100)2017-03-14
Structural Embedding of Syntactic Trees for Machine Comprehension 67.89 77.42 SEDT-LSTM2017-03-02
Bidirectional Attention Flow for Machine Comprehension✓ Link 67.777.3BIDAF (single)2016-11-05
Structural Embedding of Syntactic Trees for Machine Comprehension67.6577.19SECT-LSTM2017-03-02
Learning Recurrent Span Representations for Extractive Question Answering✓ Link66.474.9RASOR2016-11-04
Multi-Perspective Context Matching for Machine Comprehension✓ Link66.175.8MPCM2016-12-13
Dynamic Coattention Networks For Question Answering✓ Link65.475.6DCN2016-11-05
A Fully Attention-Based Information Retriever✓ Link65.175.6FABIR2018-10-22
Machine Comprehension Using Match-LSTM and Answer Pointer✓ Link64.1 64.7Match-LSTM with Bi-Ans-Ptr (Boundary+Search+b) 2016-08-29
Learning to Compute Word Embeddings On the Fly63.06OTF dict+spelling (single)2017-06-01
End-to-End Answer Chunk Extraction and Ranking for Reading Comprehension 62.571.2DCR2016-10-31
Words or Characters? Fine-grained Gating for Reading Comprehension✓ Link59.9571.25FG fine-grained gate2016-11-06
LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention✓ Link95LUKE 483M2020-10-02
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension✓ Link90.8BART Base (with text infilling)2019-10-29
Large Batch Optimization for Deep Learning: Training BERT in 76 minutes✓ Link90.584BERT large (LAMB optimizer)2019-04-01
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter✓ Link85.8DistilBERT 66M2019-10-02
Deep contextualized word representations✓ Link85.6BiDAF + Self Attention + ELMo2018-02-15