Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper Robotics Engineers,AI Researchers,Control Systems Engineers,Autonomous Systems Developers 1 week ago

Hybrid DQN-TD3 Reinforcement Learning for Autonomous Navigation in Dynamic Environments

robotics › navigation
📄 Abstract

Abstract: This paper presents a hierarchical path-planning and control framework that combines a high-level Deep Q-Network (DQN) for discrete sub-goal selection with a low-level Twin Delayed Deep Deterministic Policy Gradient (TD3) controller for continuous actuation. The high-level module selects behaviors and sub-goals; the low-level module executes smooth velocity commands. We design a practical reward shaping scheme (direction, distance, obstacle avoidance, action smoothness, collision penalty, time penalty, and progress), together with a LiDAR-based safety gate that prevents unsafe motions. The system is implemented in ROS + Gazebo (TurtleBot3) and evaluated with PathBench metrics, including success rate, collision rate, path efficiency, and re-planning efficiency, in dynamic and partially observable environments. Experiments show improved success rate and sample efficiency over single-algorithm baselines (DQN or TD3 alone) and rule-based planners, with better generalization to unseen obstacle configurations and reduced abrupt control changes. Code and evaluation scripts are available at the project repository.
Authors (4)
Xiaoyi He
Danggui Chen
Zhenshuo Zhang
Zimeng Bai
Submitted
October 30, 2025
arXiv Category
cs.RO
arXiv PDF

Key Contributions

Presents a hierarchical path-planning and control framework combining DQN for sub-goal selection and TD3 for continuous control, achieving improved success rates and sample efficiency in dynamic environments. It incorporates a practical reward shaping scheme and LiDAR-based safety.

Business Value

Enables more robust and efficient autonomous navigation for robots, reducing the need for manual intervention and improving operational reliability in complex environments.