A Comparative Analysis of Reinforcement Learning-Based Navigation for Autonomous Mobile Robot

Authors

  • Al-Mahdi Sallam Universiti Teknologi Malaysia
  • Norhaliza Abdul Wahab Universiti Teknologi Malaysia
  • Muhammad Zakiyullah Romdlony Telkom University, Bandung
  • Mohd Saiful Azimi Mahmud Universiti Teknologi Malaysia
  • Che Fai Yeong DF Automation & Robotics Sdn. Bhd

Keywords:

Reinforcement learning, Deep Q-Learning, Double Deep Q-Learning, Autonomous navigation, TurtleBot3, Gazebo simulation

Abstract

Mobile robots have been widely used in many industries including manufacturing, healthcare and warehouse automation. To ensure efficiency and safety of the robots, it is crucial to design effective control strategies that can adapt to changing environments. This paper presents reinforcement learning (RL) algorithms including Q-learning, Deep Q-Learning (DQN), and Double Deep Q-Learning (DDQN) for autonomous navigation using the Turtle-Bot3 Waffle Pi in a Gazebo-simulated environment. Three progressively complex training stages were designed to evaluate the algorithms: (1) static obstacles with predefined goals, (2) randomized goals with static obstacles, and (3) dynamic obstacles with moving goals. Performance metrics, including success rates, collision avoidance, and reward stability, were analyzed to compare algorithm effectiveness. Key results highlight DDQN’s superiority in handling complex navigation tasks. In the most challenging stage, DDQN achieved a 100% success rate and zero collisions, outperforming DQN, which attained an 88% success rate with higher collision rates. Q-learning performed well only in simple environments, as it cannot easily handle continuous state spaces. This study demonstrates the scalability of RL-based navigation systems for autonomous mobile robots. The findings provide a foundation for future advancements in dynamic and real-world robot navigation.

Author Biographies

Al-Mahdi Sallam, Universiti Teknologi Malaysia

Al Mahdi Sallam holds a Bachelor’s degree in Electrical-Mechatronics Engineering from Universiti Teknologi Malaysia (UTM). His academic and project experience centers on robotics, artificial intelligence, and automation, with a focus on reinforcement learning and AI-based inspection systems. His interests lie in intelligent control, machine learning, and real-world industrial applications of robotics.

Norhaliza Abdul Wahab, Universiti Teknologi Malaysia

Ir Dr Norhaliza Abdul Wahab is currently a Professor at Universiti Teknologi Malaysia (UTM). She completed her PhD in Electrical Engineering majoring in Control in July 2009. She is actively involved in researching and teaching in the field of industrial process control. Her expertise is in an artificial intelligent, control and automation of industrial process plant. Recently she has worked primarily on digitalization of water industry and smart manufacturing towards optimization and energy saving systems.

Muhammad Zakiyullah Romdlony, Telkom University, Bandung

Muhammad Zakiyullah Romdlony received his bachelor's and master's degrees in electrical engineering from Institut Teknologi Bandung, Indonesia, in 2009 and 2012, respectively, and a PhD degree in systems & control from the University of Groningen, the Netherlands, in 2018. He is an associate professor at the School of Electrical Engineering, Telkom University, Indonesia. His research interests include safety-critical systems, robust control, and robotics.

Mohd Saiful Azimi Mahmud, Universiti Teknologi Malaysia

Dr. Mohd Saiful Azimi Mahmud is a Senior Lecturer at the Department of Control and Mechatronics Engineering, Faculty of Electrical Engineering, Universiti Teknologi Malaysia (UTM). He earned his B.Eng. (Electrical- Mechatronics) from UTM in 2015. Driven by his passion for robotics, he pursued his doctoral studies in Electrical Engineering, focusing on robot navigation, and completed his PhD in April 2019 at UTM. His research interests include multi-objective optimization, robot navigation, artificial intelligence, robot and machine scheduling, and smart manufacturing. His expertise has led him to secure several internal and national research grants. In recognition of his contributions, he was honored with the Young Scientist Award at an international conference held in Kyoto, Japan in 2025.

Che Fai Yeong , DF Automation & Robotics Sdn. Bhd

Dr. Yeong Che Fai is the Chief Executive Officer of DF Automation & Robotics Sdn. Bhd. He obtained his Ph.D. from Imperial College London and earned both his Master of Engineering and Bachelor of Engineering degrees from Universiti Teknologi Malaysia (UTM). Prior to his current role, he served as an Associate Professor at UTM. His areas of expertise and interest include artificial intelligence, robotics, and smart manufacturing.

References

[1] Q. Zou, Q. Sun, L. Chen, B. Nie and Q. Li, A comparative analysis of lidar slam-based indoor navigation for autonomous vehicles, IEEE Transactions on Intelligent Transportation Systems, 23, 202, 6907-6921.

[2] D. Küpper, L. Markus, K. Claudio, K. Kristian, M. Andreas, R. Lässig and T. Buchner, Advanced Robotics in the Factory of the Future, Boston Consulting Group, 2021.

[3] L. Chen, W. Zhan, W. Tian, Y. He and Q. Zou, Deep integration: A multi-label architecture for road scene recognition, IEEE Transactions on Image Processing, 28(10), 2019, 4883-4898.

[4] S. Kumar, V. Manjrekar, V. Singh and B. Kumar, Integrated yet distributed operations planning approach: A next generation manufacturing planning system, Journal of Manufacturing Systems, 54, 2020, 103-122.

[5] M. B. Alatise and G. P. Hancke, A review on challenges of autonomous mobile robot and sensor fusion methods, IEEE Access, 8, 2020, 39830-39846.

[6] V. R. F. Miranda, A. A. Neto, G. M. Freitas and L. A. Mozelli, Generalization in deep reinforcement learning for robotic navigation by reward shaping, IEEE Transactions on Industrial Electronics, 71, 2024, 6013-6020.

[7] J. Pak, J. Kim, Y. Park and H. I. Son, Field evaluation of path-planning algorithms for autonomous mobile robots in smart farms, IEEE Access, 10, 2022, 60253-60266.

[8] K. T. Song, Y. H. Chiu, L. R. Kang, S. H. Song, C. A. Yang, P. C. Lu and S. Q. Ou, Navigation control design of a mobile robot by integrating obstacle avoidance and lidar slam, Proceedings of 2018 IEEE International Conference on Systems, Man, and Cybernetics, Miyazaki, Japan, 2018, 1833-1838.

[9] M. N. A. Wahab, S. Nefti-Meziani and A. Atyabi, A comparative review on mobile robot path planning: Classical or meta-heuristic methods?, Annual Reviews in Control, 50, 2020, 233-252.

[10] S. Kadry, G. Alferov, V. Fedorov and A. Khokhriakova, Path optimization for D-Star algorithm modification, AIP Conference Proceedings, 2425(1), 2022.

[11] J. S. Smith, R. Xu and P. Vela, EgoTEB: Egocentric, perception space navigation using timed-elastic-bands, Proceedings of IEEE International Conference on Robotics and Automation, Paris, France, 2020, 2703-2709.

[12] J. Zhao, S. Liu and J. Li, Research and implementation of autonomous navigation for mobile robots based on SLAM algorithm under ROS, Sensors, 22(11), 2022, 4172.

[13] S. M. B. P. Samarakoon, M. A. V. J. Muthugala and M. R. Elara, Global and local area coverage path planner for a reconfigurable robot, Proceedings of 2022 IEEE Congress on Evolutionary Computation, Padua, Italy, 2022, 1-8.

[14] Q. Feng, K. Du, H. Wang, T. Chen, X. Meng, S. Wang and B. Lu, Path planning algorithm of mobile robot based on improved q-learning algorithm, Proceedings of 2023 IEEE Information Technology, Networking, Electronic and Automation Control Conference, Chongqing, China, 2023, 133-136.

[15] S. W. H. Teoh, K. Kamarudin, N. A. N. Ali, M. M. M. Zainal, M. R. Manan and S. M. Mamduh, Reinforcement learning for mobile robot's environment exploration, Journal of Physics: Conference Series, 2641, 2023, 012003.

[16] R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction, 2nd ed., Cambridge, MA: MIT Press, 2018.

[17] K. Zhu and T. Zhang, Deep reinforcement learning-based mobile robot navigation: A review, Tsinghua Science and Technology, 26, 2021, 674-691.

[18] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare and D. Hassabis, Human-level control through deep reinforcement learning, Nature, 518, 2015, 529-533.

[19] Y. Yuxiang, W. Shuting, X. Yuanlong, H. Yiming and L. Hu, Q-learning-based collision-free path planning for mobile robot in unknown environment, Proceedings of 2022 IEEE International Conference on Industrial Electronics and Applications, Chengdu, China, 2022, 1104-1109.

[20] C. Vishal and J. Amudha, Autonomous driving mobile robot using q-learning, Proceedings of 2022 IEEE International Conference on Frontiers of Technology, Belgaum, India, 2022, 1-8.

[21] Y. Wang and S. Yang, Navigation for mobile robots using reinforcement learning and dynamic obstacle detection, IEEE Access, 10, 2022, 18001-18010.

[22] H. Hasselt, A. Guez and D. Silver, Deep reinforcement learning with double q-learning, Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, Arizona, USA, 30(1), 2016.

[23] L. Tai, G. Paolo and M. Liu, Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation, Proceedings of 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vancouver, Canada, 2017, 31-36.

[24] Z. Yue, Z. Liu, Y. Miao and H. Wang, Efficient reinforcement learning for 3d lidar navigation of mobile robots, Proceedings of Chinese Control Conference, Hefei, China, 2022, 3755-3760.

Downloads

Published

26-11-2025

How to Cite

Sallam, A.-M., Abdul Wahab, N., Romdlony, M. Z., Mahmud, M. S. A., & Yeong , C. F. (2025). A Comparative Analysis of Reinforcement Learning-Based Navigation for Autonomous Mobile Robot. Applications of Modelling and Simulation, 9, 389–403. Retrieved from https://www.ojs.arqiipubl.com/index.php/AMS_Journal/article/view/1154

Issue

Section

Articles

Most read articles by the same author(s)

Similar Articles

<< < 1 2 3 4 5 6 

You may also start an advanced similarity search for this article.