OpenCodePapers

question-answering-on-storycloze

Question Answering
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAccuracyModelNameReleaseDate
Crosslingual Generalization through Multitask Finetuning✓ Link96.3BLOOMZ2022-11-03
Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners✓ Link95.88Flipped-3B2022-10-06
Finetuned Language Models Are Zero-Shot Learners✓ Link94.7FLAN 137B (few-shot, k=10)2021-09-03
The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning✓ Link94.5T0-3B (CoT fine-tuned)2023-05-23
Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models94.40KiC-770M2022-10-28
Finetuned Language Models Are Zero-Shot Learners✓ Link93.4FLAN 137B (zero-shot)2021-09-03
Improving Machine Reading Comprehension with General Reading Strategies✓ Link88.3Reading Strategies Model2018-10-31
Improving Language Understanding by Generative Pre-Training✓ Link86.5Finetuned Transformer LM2018-06-11
Exploring the Benefits of Training Expert Language Models over Instruction Tuning✓ Link86.33RoE-3B2023-02-07
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot✓ Link79.82OPT-175B2023-01-02
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot✓ Link78.87SparseGPT (175B, 50% Sparsity)2023-01-02
UNIMELB at SemEval-2016 Tasks 4A and 4B: An Ensemble of Neural Networks and a Word2Vec Based Model for Sentiment Classification✓ Link78.7Memory chains and semantic supervision2016-06-01
Story Comprehension for Predicting What Happens Next77.6Hidden Coherence Model2017-09-01
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot✓ Link77.02SparseGPT (175B, 4:8 Sparsity)2023-01-02
A Simple and Effective Approach to the Story Cloze Test76.5val-LS-skip2018-03-15
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot✓ Link76.19SparseGPT (175B, 2:4 Sparsity)2023-01-02
Efficient Language Modeling with Sparse all-MLP74.7sMLP – deterministic 9.4B (0-shot)2022-03-14
Efficient Language Modeling with Sparse all-MLP73.3Switch Transformer 9B2022-03-14
Language Models are Few-Shot Learners✓ Link72.4GPT-3 Large 760M (zero-shot)2020-05-28
Efficient Language Modeling with Sparse all-MLP67.9Gshard 9B2022-03-14
Efficient Language Modeling with Sparse all-MLP64.7HASH Layers 10B (0-shot)2022-03-14
Efficient Language Modeling with Sparse all-MLP61.4Base Layers 10B (0-shot)2022-03-14
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot✓ Link47.10OPT-175B (50% Sparsity)2023-01-02