OpenCodePapers

efficient-vits-on-imagenet-1k-with-deit-t

Image ClassificationEfficient ViTs
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeTop 1 AccuracyGFLOPsModelNameReleaseDate
Joint Token Pruning and Squeezing Towards More Aggressive Compression of Vision Transformers✓ Link72.90.8dTPS2023-04-21
Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers✓ Link72.91.0MCTF ($r=8$)2024-03-15
Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers✓ Link72.70.7MCTF ($r=16$)2024-03-15
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link72.30.8BAT2022-11-21
Joint Token Pruning and Squeezing Towards More Aggressive Compression of Vision Transformers✓ Link72.30.8eTPS2023-04-21
SPViT: Enabling Faster Vision Transformers via Soft Token Pruning✓ Link72.21.0SPViT (1.0G)2021-12-27
Training data-efficient image transformers & distillation through attention✓ Link72.21.2Base (DeiT-T)2020-12-23
Patch Slimming for Efficient Vision Transformers72.10.6DPS-ViT2021-06-05
PPT: Token Pruning and Pooling for Efficient Vision Transformers✓ Link72.10.8PPT2023-10-03
SPViT: Enabling Faster Vision Transformers via Soft Token Pruning✓ Link72.10.9SPViT (0.9G)2021-12-27
Patch Slimming for Efficient Vision Transformers72.00.7PS-ViT2021-06-05
Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer✓ Link72.00.8EvoViT2021-08-03
Learned Thresholds Token Merging and Pruning for Vision Transformers✓ Link72.01.0LTMP (80%)2023-07-20
Token Merging: Your ViT But Faster✓ Link71.70.9ToMe ($r=8$)2022-10-17
Learned Thresholds Token Merging and Pruning for Vision Transformers✓ Link71.50.8LTMP (60%)2023-07-20
Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers✓ Link71.40.6MCTF ($r=20$)2024-03-15
Token Merging: Your ViT But Faster✓ Link71.40.8ToMe ($r=12$)2022-10-17
Token Merging: Your ViT But Faster✓ Link70.70.6ToMe ($r=16$)2022-10-17
Pruning Self-attentions into Convolutional Layers in Single Path✓ Link70.71.0SPViT2021-11-23
Chasing Sparsity in Vision Transformers: An End-to-End Exploration✓ Link70.10.9S$^2$ViTE2021-06-08
Learned Thresholds Token Merging and Pruning for Vision Transformers✓ Link69.80.7LTMP (45%)2023-07-20
Scalable Vision Transformers with Hierarchical Pooling✓ Link69.60.6HVT-Ti-12021-03-19