A Comparative Analysis of Reinforcement Learning-Based Navigation for Autonomous Mobile Robot
Keywords:
Reinforcement learning, Deep Q-Learning, Double Deep Q-Learning, Autonomous navigation, TurtleBot3, Gazebo simulationAbstract
Mobile robots have been widely used in many industries including manufacturing, healthcare and warehouse automation. To ensure efficiency and safety of the robots, it is crucial to design effective control strategies that can adapt to changing environments. This paper presents reinforcement learning (RL) algorithms including Q-learning, Deep Q-Learning (DQN), and Double Deep Q-Learning (DDQN) for autonomous navigation using the Turtle-Bot3 Waffle Pi in a Gazebo-simulated environment. Three progressively complex training stages were designed to evaluate the algorithms: (1) static obstacles with predefined goals, (2) randomized goals with static obstacles, and (3) dynamic obstacles with moving goals. Performance metrics, including success rates, collision avoidance, and reward stability, were analyzed to compare algorithm effectiveness. Key results highlight DDQN’s superiority in handling complex navigation tasks. In the most challenging stage, DDQN achieved a 100% success rate and zero collisions, outperforming DQN, which attained an 88% success rate with higher collision rates. Q-learning performed well only in simple environments, as it cannot easily handle continuous state spaces. This study demonstrates the scalability of RL-based navigation systems for autonomous mobile robots. The findings provide a foundation for future advancements in dynamic and real-world robot navigation.References
[1] Q. Zou, Q. Sun, L. Chen, B. Nie and Q. Li, A comparative analysis of lidar slam-based indoor navigation for autonomous vehicles, IEEE Transactions on Intelligent Transportation Systems, 23, 202, 6907-6921.
[2] D. Küpper, L. Markus, K. Claudio, K. Kristian, M. Andreas, R. Lässig and T. Buchner, Advanced Robotics in the Factory of the Future, Boston Consulting Group, 2021.
[3] L. Chen, W. Zhan, W. Tian, Y. He and Q. Zou, Deep integration: A multi-label architecture for road scene recognition, IEEE Transactions on Image Processing, 28(10), 2019, 4883-4898.
[4] S. Kumar, V. Manjrekar, V. Singh and B. Kumar, Integrated yet distributed operations planning approach: A next generation manufacturing planning system, Journal of Manufacturing Systems, 54, 2020, 103-122.
[5] M. B. Alatise and G. P. Hancke, A review on challenges of autonomous mobile robot and sensor fusion methods, IEEE Access, 8, 2020, 39830-39846.
[6] V. R. F. Miranda, A. A. Neto, G. M. Freitas and L. A. Mozelli, Generalization in deep reinforcement learning for robotic navigation by reward shaping, IEEE Transactions on Industrial Electronics, 71, 2024, 6013-6020.
[7] J. Pak, J. Kim, Y. Park and H. I. Son, Field evaluation of path-planning algorithms for autonomous mobile robots in smart farms, IEEE Access, 10, 2022, 60253-60266.
[8] K. T. Song, Y. H. Chiu, L. R. Kang, S. H. Song, C. A. Yang, P. C. Lu and S. Q. Ou, Navigation control design of a mobile robot by integrating obstacle avoidance and lidar slam, Proceedings of 2018 IEEE International Conference on Systems, Man, and Cybernetics, Miyazaki, Japan, 2018, 1833-1838.
[9] M. N. A. Wahab, S. Nefti-Meziani and A. Atyabi, A comparative review on mobile robot path planning: Classical or meta-heuristic methods?, Annual Reviews in Control, 50, 2020, 233-252.
[10] S. Kadry, G. Alferov, V. Fedorov and A. Khokhriakova, Path optimization for D-Star algorithm modification, AIP Conference Proceedings, 2425(1), 2022.
[11] J. S. Smith, R. Xu and P. Vela, EgoTEB: Egocentric, perception space navigation using timed-elastic-bands, Proceedings of IEEE International Conference on Robotics and Automation, Paris, France, 2020, 2703-2709.
[12] J. Zhao, S. Liu and J. Li, Research and implementation of autonomous navigation for mobile robots based on SLAM algorithm under ROS, Sensors, 22(11), 2022, 4172.
[13] S. M. B. P. Samarakoon, M. A. V. J. Muthugala and M. R. Elara, Global and local area coverage path planner for a reconfigurable robot, Proceedings of 2022 IEEE Congress on Evolutionary Computation, Padua, Italy, 2022, 1-8.
[14] Q. Feng, K. Du, H. Wang, T. Chen, X. Meng, S. Wang and B. Lu, Path planning algorithm of mobile robot based on improved q-learning algorithm, Proceedings of 2023 IEEE Information Technology, Networking, Electronic and Automation Control Conference, Chongqing, China, 2023, 133-136.
[15] S. W. H. Teoh, K. Kamarudin, N. A. N. Ali, M. M. M. Zainal, M. R. Manan and S. M. Mamduh, Reinforcement learning for mobile robot's environment exploration, Journal of Physics: Conference Series, 2641, 2023, 012003.
[16] R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction, 2nd ed., Cambridge, MA: MIT Press, 2018.
[17] K. Zhu and T. Zhang, Deep reinforcement learning-based mobile robot navigation: A review, Tsinghua Science and Technology, 26, 2021, 674-691.
[18] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare and D. Hassabis, Human-level control through deep reinforcement learning, Nature, 518, 2015, 529-533.
[19] Y. Yuxiang, W. Shuting, X. Yuanlong, H. Yiming and L. Hu, Q-learning-based collision-free path planning for mobile robot in unknown environment, Proceedings of 2022 IEEE International Conference on Industrial Electronics and Applications, Chengdu, China, 2022, 1104-1109.
[20] C. Vishal and J. Amudha, Autonomous driving mobile robot using q-learning, Proceedings of 2022 IEEE International Conference on Frontiers of Technology, Belgaum, India, 2022, 1-8.
[21] Y. Wang and S. Yang, Navigation for mobile robots using reinforcement learning and dynamic obstacle detection, IEEE Access, 10, 2022, 18001-18010.
[22] H. Hasselt, A. Guez and D. Silver, Deep reinforcement learning with double q-learning, Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, Arizona, USA, 30(1), 2016.
[23] L. Tai, G. Paolo and M. Liu, Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation, Proceedings of 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vancouver, Canada, 2017, 31-36.
[24] Z. Yue, Z. Liu, Y. Miao and H. Wang, Efficient reinforcement learning for 3d lidar navigation of mobile robots, Proceedings of Chinese Control Conference, Hefei, China, 2022, 3755-3760.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Al-Mahdi Sallam, Norhaliza Abdul Wahab, Muhammad Zakiyullah Romdlony, Mohd Saiful Azimi Mahmud, Che Fai Yeong

This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors who publish with this journal agree to the following terms:Authors hold and retain copyright, and grant the journal right of first publication, with the work after publication simultaneously licensed under a Creative Commons Attribution 4.0 License CC BY that permits any use, reproduction and distribution of the work and article without further permission provided that the original work is properly cited.
Authors are permitted and encouraged to post their work online in institutional repositories, website and other social media before and after publication, as it can lead to productive exchanges, as well as earlier and greater citation of published work.





