Paper | Code | Average Accuracy | Character Error Rate (CER) | Word Error Rate (WER) | ModelName | ReleaseDate |
---|---|---|---|---|---|---|
Benchmarking Vision-Language Models on Optical Character Recognition in Dynamic Video Environments | ✓ Link | 76.22 | 0.2378 | 0.5117 | GPT-4o | 2025-02-10 |
Benchmarking Vision-Language Models on Optical Character Recognition in Dynamic Video Environments | ✓ Link | 76.13 | 0.2387 | 0.2385 | Gemini-1.5 Pro | 2025-02-10 |
Benchmarking Vision-Language Models on Optical Character Recognition in Dynamic Video Environments | ✓ Link | 67.71 | 0.3229 | 0.4663 | Claude-3 Sonnet | 2025-02-10 |
Benchmarking Vision-Language Models on Optical Character Recognition in Dynamic Video Environments | ✓ Link | 56.98 | 0.7620 | 0.4302 | RapidOCR | 2025-02-10 |
Benchmarking Vision-Language Models on Optical Character Recognition in Dynamic Video Environments | ✓ Link | 49.30 | 0.5070 | 0.8262 | EasyOCR | 2025-02-10 |