OpenCodePapers

natural-language-inference-on-rcb

Natural Language Inference
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeAverage F1AccuracyModelNameReleaseDate
RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark✓ Link0.680.702Human Benchmark2020-10-29
[]()0.4520.484RuBERT conversational
[]()0.417 0.484RuGPT3Large
[]()0.4080.447YaLM 1.0B few-shot
[]()0.4060.546Golden Transformer
Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks0.40.438heuristic majority2021-05-03
[]()0.3720.461RuGPT3Medium
[]()0.3710.452SBERT_Large
[]()0.3670.463RuBERT plain
[]()0.3670.445Multilingual Bert
mT5: A massively multilingual pre-trained text-to-text transformer✓ Link0.3660.454MT5 Large2020-10-22
[]()0.3570.518ruRoberta-large finetune
[]()0.3560.5ruBert-large finetune
[]()0.3560.473RuGPT3Small
[]()0.3510.486SBERT_Large_mt_ru_finetuning
[]()0.3330.509ruBert-base finetune
Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks0.3190.374Random weighted2021-05-03
[]()0.3070.468ruT5-base-finetune
[]()0.3060.498ruT5-large-finetune
[]()0.3020.418RuGPT3XL few-shot
RussianSuperGLUE: A Russian Language Understanding Evaluation Benchmark✓ Link0.3010.441Baseline TF-IDF1.12020-10-29
Unreasonable Effectiveness of Rule-Based Heuristics in Solving Russian SuperGLUE Tasks0.2170.484majority_class2021-05-03