OpenCodePapers

efficient-vits-on-imagenet-1k-with-deit-s

Image ClassificationEfficient ViTs
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodeTop 1 AccuracyGFLOPsModelNameReleaseDate
Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers✓ Link80.12.6MCTF ($r=16$)2024-03-15
Joint Token Pruning and Squeezing Towards More Aggressive Compression of Vision Transformers✓ Link80.13.0dTPS2023-04-21
Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers✓ Link79.92.4MCTF ($r=18$)2024-03-15
DiffRate : Differentiable Compression Rate for Efficient Vision Transformers✓ Link79.82.9DiffRate2023-05-29
PPT: Token Pruning and Pooling for Efficient Vision Transformers✓ Link79.82.9PPT2023-10-03
DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification✓ Link79.83.4DynamicViT (80%)2021-06-03
Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations✓ Link79.83.5EViT (80%)2022-02-16
Learned Thresholds Token Merging and Pruning for Vision Transformers✓ Link79.83.8LTMP (80%)2023-07-20
SPViT: Enabling Faster Vision Transformers via Soft Token Pruning✓ Link79.83.9SPViT (3.9G)2021-12-27
Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations✓ Link79.84.0EViT (90%)2022-02-16
DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification✓ Link79.84.0DynamicViT (90%)2021-06-03
Training data-efficient image transformers & distillation through attention✓ Link79.84.6Base (DeiT-S)2020-12-23
Adaptive Token Sampling For Efficient Vision Transformers✓ Link79.72.9ATS2021-11-30
Joint Token Pruning and Squeezing Towards More Aggressive Compression of Vision Transformers✓ Link79.73.0eTPS2023-04-21
Token Merging: Your ViT But Faster✓ Link79.73.4ToMe ($r=8$)2022-10-17
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link79.63.0BAT (70%)2022-11-21
Adaptive Sparse ViT: Towards Learnable Adaptive Token Pruning by Fully Exploiting Self-Attention✓ Link79.63.0AS-DeiT-S (65%)2022-09-28
Learned Thresholds Token Merging and Pruning for Vision Transformers✓ Link79.63.0LTMP (60%)2023-07-20
Multi-criteria Token Fusion with One-step-ahead Attention for Efficient Vision Transformers✓ Link79.52.2MCTF ($r=20$)2024-03-15
Patch Slimming for Efficient Vision Transformers79.52.4DPS-ViT2021-06-05
Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations✓ Link79.53.0EViT (70%)2022-02-16
Patch Slimming for Efficient Vision Transformers79.42.6PS-ViT2021-06-05
Token Merging: Your ViT But Faster✓ Link79.42.7ToMe ($r=13$)2022-10-17
Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer✓ Link79.43.0EvoViT2021-08-03
SPViT: Enabling Faster Vision Transformers via Soft Token Pruning✓ Link79.32.6SPViT (2.6G)2021-12-27
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link79.32.6BAT (60%)2022-11-21
DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification✓ Link79.32.9DynamicViT (70%)2021-06-03
Chasing Sparsity in Vision Transformers: An End-to-End Exploration✓ Link79.23.2S$^2$ViTE2021-06-08
Token Merging: Your ViT But Faster✓ Link79.12.3ToMe ($r=16$)2022-10-17
IA-RED$^2$: Interpretability-Aware Redundancy Reduction for Vision Transformers79.13.2IA-RED$^2$2021-06-23
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link79.02.3BAT (50%)2022-11-21
Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations✓ Link78.92.6EViT (60%)2022-02-16
Adaptive Sparse ViT: Towards Learnable Adaptive Token Pruning by Fully Exploiting Self-Attention✓ Link78.72.3AS-DeiT-S (50%)2022-09-28
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link78.62.0BAT (40%)2022-11-21
Learned Thresholds Token Merging and Pruning for Vision Transformers✓ Link78.62.3LTMP (45%)2023-07-20
AdaViT: Adaptive Tokens for Efficient Vision Transformer✓ Link78.63.6A-ViT2021-12-14
Not All Patches are What You Need: Expediting Vision Transformers via Token Reorganizations✓ Link78.52.3EViT (50%)2022-02-16
Scalable Vision Transformers with Hierarchical Pooling✓ Link78.32.7HVT-S-12021-03-19
Pruning Self-attentions into Convolutional Layers in Single Path✓ Link78.33.3SPViT2021-11-23
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link77.81.8BAT (30%)2022-11-21
Beyond Attentive Tokens: Incorporating Token Importance and Diversity for Efficient Vision Transformers✓ Link76.41.6BAT (20%)2022-11-21