OpenCodePapers

visual-navigation-on-room-to-room-1

Visual Navigation
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
PaperCodesplModelNameReleaseDate
Agent Journey Beyond RGB: Unveiling Hybrid Semantic-Spatial Environmental Representations for Vision-and-Language Navigation✓ Link0.6383SUSA2024-12-09
Meta-Explore: Exploratory Hierarchical Vision-and-Language Navigation Using Scene Object Spectrum Grounding0.61Meta-Explore2023-03-07
BEVBert: Multimodal Map Pre-training for Language-guided Navigation✓ Link0.60BEV-BERT2022-12-08
Towards Learning a Generalist Model for Embodied Navigation✓ Link0.60NaviLLM2023-12-04
HOP: History-and-Order Aware Pre-training for Vision-and-Language Navigation✓ Link0.59HOP2022-03-22
VLN-PETL: Parameter-Efficient Transfer Learning for Vision-and-Language Navigation✓ Link0.58VLN-PETL2023-08-20
Think Global, Act Local: Dual-scale Graph Transformer for Vision-and-Language Navigation✓ Link0.58DUET2022-02-23
A Recurrent Vision-and-Language BERT for Navigation✓ Link0.57VLN-BERT2020-11-26
Towards Learning a Generic Agent for Vision-and-Language Navigation via Pre-training✓ Link0.51Prevalent2020-02-25
Reinforced Cross-Modal Matching and Self-Supervised Imitation Learning for Vision-Language Navigation0.38RCM+SIL(no early exploration)2018-11-25
Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments✓ Link0.18Seq2Seq baseline2017-11-20