OpenCodePapers

domain-generalization-on-imagenet-r

Domain Generalization
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeTop-1 Error RateModelNameReleaseDate
Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time✓ Link3.90Model soups (BASIC-L)2022-03-10
Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time✓ Link4.54Model soups (ViT-G/14)2022-03-10
Context-Aware Robust Fine-Tuning10.3CAR-FT (CLIP, ViT-L/14@336px)2022-11-29
Understanding The Robustness in Vision Transformers✓ Link28.9FAN-Hybrid-L(IN-21K, 384))2022-04-26
MetaFormer Baselines for Vision✓ Link29.6CAFormer-B36 (IN21K, 384)2022-10-24
A Whac-A-Mole Dilemma: Shortcuts Come in Multiples Where Mitigating One Amplifies Others✓ Link31.3LLE (ViT-B/16, SWAG, Edge Aug)2022-12-09
MetaFormer Baselines for Vision✓ Link31.7CAFormer-B36 (IN21K)2022-10-24
A ConvNet for the 2020s✓ Link31.8ConvNeXt-XL (Im21k, 384)2022-01-10
A Whac-A-Mole Dilemma: Shortcuts Come in Multiples Where Mitigating One Amplifies Others✓ Link33.1LLE (ViT-H/14, MAE, Edge Aug)2022-12-09
Masked Autoencoders Are Scalable Vision Learners✓ Link33.5MAE (ViT-H, 448)2021-11-11
MetaFormer Baselines for Vision✓ Link33.5ConvFormer-B36 (IN21K, 384)2022-10-24
Enhance the Visual Representation via Discrete Adversarial Training✓ Link34.39MAE+DAT (ViT-H)2022-09-16
MetaFormer Baselines for Vision✓ Link34.7ConvFormer-B36 (IN21K)2022-10-24
Distilling Out-of-Distribution Robustness from Vision-Language Foundation Models✓ Link34.9Discrete Adversarial Distillation (ViT-B,224)2023-11-02
Generalized Parametric Contrastive Learning✓ Link39.7GPaCo (ViT-L)2022-09-26
Improving Vision Transformers by Revisiting High-frequency Components✓ Link40.3VOLO-D5+HAT2022-04-03
Pyramid Adversarial Training Improves ViT Performance✓ Link42.16Pyramid Adversarial Training Improves ViT (Im21k)2021-11-30
Fully Attentional Networks with Self-emerging Token Labeling✓ Link43.4FAN-L-Hybrid+STL2024-01-08
Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without Supervision✓ Link43.9SEER (RegNet10B)2022-02-16
Discrete Representations Strengthen Vision Transformer Robustness✓ Link44.74DiscreteViT2021-11-20
MetaFormer Baselines for Vision✓ Link45CAFormer-B36 (384)2022-10-24
Pyramid Adversarial Training Improves ViT Performance✓ Link46.08Pyramid Adversarial Training Improves ViT2021-11-30
MetaFormer Baselines for Vision✓ Link46.1CAFormer-B362022-10-24
MetaFormer Baselines for Vision✓ Link47.8ConvFormer-B36 (384)2022-10-24
MetaFormer Baselines for Vision✓ Link48.9ConvFormer-B362022-10-24
Towards Robust Vision Transformer✓ Link51.3RVT-B*2021-05-17
Sequencer: Deep LSTM for Image Classification✓ Link51.9Sequencer2D-L2022-05-04
Towards Robust Vision Transformer✓ Link52.3RVT-S*2021-05-17
The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization✓ Link 53.2 DeepAugment+AugMix (ResNet-50)2020-06-29
PRIME: A few primitives can boost robustness to common corruptions✓ Link53.7PRIME with JSD (ResNet-50)2021-12-27
Towards Robust Vision Transformer✓ Link56.1RVT-Ti*2021-05-17
PRIME: A few primitives can boost robustness to common corruptions✓ Link57.1PRIME (ResNet-50)2021-12-27
The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization✓ Link57.8DeepAugment (ResNet-50)2020-06-29
ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness✓ Link58.5 Stylized ImageNet (ResNet-50)2018-11-29
AugMix: A Simple Data Processing Method to Improve Robustness and Uncertainty✓ Link58.9AugMix (ResNet-50)2019-12-05
Deep Residual Learning for Image Recognition✓ Link63.9ResNet-502015-12-10
When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations✓ Link71.9ResNet-152x2-SAM2021-06-03
When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations✓ Link73.6ViT-B/16-SAM2021-06-03
When Vision Transformers Outperform ResNets without Pre-training or Strong Data Augmentations✓ Link76.5Mixer-B/8-SAM2021-06-03