Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 97% Match Research Paper Robotics researchers,AI developers,Embodied AI researchers,Robotics engineers 1 day ago

Fast-SmartWay: Panoramic-Free End-to-End Zero-Shot Vision-and-Language Navigation

robotics › navigation
📄 Abstract

Abstract: Recent advances in Vision-and-Language Navigation in Continuous Environments (VLN-CE) have leveraged multimodal large language models (MLLMs) to achieve zero-shot navigation. However, existing methods often rely on panoramic observations and two-stage pipelines involving waypoint predictors, which introduce significant latency and limit real-world applicability. In this work, we propose Fast-SmartWay, an end-to-end zero-shot VLN-CE framework that eliminates the need for panoramic views and waypoint predictors. Our approach uses only three frontal RGB-D images combined with natural language instructions, enabling MLLMs to directly predict actions. To enhance decision robustness, we introduce an Uncertainty-Aware Reasoning module that integrates (i) a Disambiguation Module for avoiding local optima, and (ii) a Future-Past Bidirectional Reasoning mechanism for globally coherent planning. Experiments on both simulated and real-robot environments demonstrate that our method significantly reduces per-step latency while achieving competitive or superior performance compared to panoramic-view baselines. These results demonstrate the practicality and effectiveness of Fast-SmartWay for real-world zero-shot embodied navigation.
Authors (4)
Xiangyu Shi
Zerui Li
Yanyuan Qiao
Qi Wu
Submitted
November 2, 2025
arXiv Category
cs.RO
arXiv PDF

Key Contributions

Fast-SmartWay presents an end-to-end zero-shot Vision-and-Language Navigation framework that eliminates the need for panoramic views and waypoint predictors, using only three frontal RGB-D images. It enhances decision robustness with an Uncertainty-Aware Reasoning module, enabling MLLMs to directly predict actions and achieve significantly improved performance in both simulated and real-robot environments.

Business Value

Enables more responsive and adaptable robots for tasks like indoor navigation, delivery, and assistance, reducing development complexity and improving user experience.