OpenCodePapers

natural-language-inference-on-terra

Natural Language Inference
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAccuracyModelNameReleaseDate
RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark✓ Link0.92Human Benchmark2020-10-29
[]()0.871Golden Transformer
[]()0.801ruRoberta-large finetune
[]()0.747ruT5-large-finetune
[]()0.704ruBert-large finetune
[]()0.703ruBert-base finetune
[]()0.692ruT5-base-finetune
[]()0.654RuGPT3Large
[]()0.642RuBERT plain
[]()0.64RuBERT conversational
[]()0.637SBERT_Large_mt_ru_finetuning
[]()0.637SBERT_Large
[]()0.617Multilingual Bert
[]()0.605YaLM 1.0B few-shot
[]()0.573RuGPT3XL few-shot
mT5: A massively multilingual pre-trained text-to-text transformer✓ Link0.561MT5 Large2020-10-22
Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks0.549heuristic majority2021-05-03
Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks0.513majority_class2021-05-03
[]()0.505RuGPT3Medium
[]()0.488RuGPT3Small
Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks0.483Random weighted2021-05-03
RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark✓ Link0.471Baseline TF-IDF1.12020-10-29