OpenCodePapers
vision-and-language-navigation-on-touchdown
Vision and Language Navigation
Dataset Link
Results over time
Click legend items to toggle metrics. Hover points for model names.
Leaderboard
Show papers without code
Paper
Code
Task Completion (TC)
↕
ModelName
ReleaseDate
↕
FLAME: Learning to Navigate with Multimodal LLM in Urban Environments
✓ Link
40.20
FLAME
2024-08-20
Analyzing Generalization of Vision and Language Navigation to Unseen Outdoor Areas
✓ Link
29.1
ORAR + junction type + heading delta
2022-03-25
Analyzing Generalization of Vision and Language Navigation to Unseen Outdoor Areas
✓ Link
24.2
ORAR
2022-03-25
Learning to Stop: A Simple yet Effective Approach to Urban Vision-Language Navigation
16.68
ARC + L2STOP
2020-09-28
Multimodal Text Style Transfer for Outdoor Vision-and-Language Navigation
✓ Link
16.2
VLN Transformer +M-50 +style
2020-07-01
Multimodal Text Style Transfer for Outdoor Vision-and-Language Navigation
✓ Link
14.9
VLN Transformer
2020-07-01
Learning to Stop: A Simple yet Effective Approach to Urban Vision-Language Navigation
14.13
ARC
2020-09-28
Retouchdown: Adding Touchdown to StreetLearn as a Shareable Resource for Language Grounding Tasks in Street View
✓ Link
12.8
Retouch-RConcat
2020-01-10
Multimodal Text Style Transfer for Outdoor Vision-and-Language Navigation
✓ Link
11.9
Gated Attention (GA)
2020-07-01
Multimodal Text Style Transfer for Outdoor Vision-and-Language Navigation
✓ Link
11.8
RConcat
2020-07-01
Touchdown: Natural Language Navigation and Spatial Reasoning in Visual Street Environments
✓ Link
10.7
RConcat
2018-11-29
Touchdown: Natural Language Navigation and Spatial Reasoning in Visual Street Environments
✓ Link
5.5
Gated Attention (GA)
2018-11-29