METHOD OF LEARNING OF AUTONOMOUS MOBILE ROBOTS BASED ON DRL AND CURRICULUM LEARNING

Authors

DOI:

https://doi.org/10.28925/2663-4023.2025.30.994

Keywords:

information technology, machine learning methods, reinforcement learning methods, deep reinforcement learning, curriculum learning, autonomous mobile robots, mobile robot navigation, ROS 2, Gazebo.

Abstract

The work is devoted to the urgent task of improving the efficiency of socially adaptive navigation of autonomous mobile robots in dynamic environments with human presence. The application of deep reinforcement learning (DRL) methods to solve this problem is complicated by the high dimensionality of the state space, the complexity of formalizing social norms in the reward function, and the instability of the learning process. To overcome these challenges, a method is proposed that integrates the Proximal Policy Optimization (PPO) algorithm with the Curriculum Learning (CL) training strategy. The developed training program combines a gradual increase in the complexity of the environment (from static obstacles to an environment with moving human agents) and the phased formation of the reward function with the addition of social components. A key feature is the transition between stages, which is based on policy stability analysis. The experimental study was conducted in the developed Gazebo simulation environment using the Turtlebot3 Waffle mobile robot and the ROS 2 Humble framework. Step-by-step training allows an autonomous mobile robot to first learn basic skills for avoiding static obstacles, then dynamic ones, and finally, at the final stage, to take into account social norms of interaction with people. The input data for the system is data from LiDAR, the status of the robot and people, and the target position. The result of the method is an optimized stochastic behavior policy that allows an autonomous mobile robot to make safe, efficient, and socially acceptable navigation decisions. A comparative analysis of the proposed method with the standard PPO algorithm was performed. The results confirm that the proposed method allows the formation of an effective policy of socially adaptive navigation, solving the problems of instability and slow convergence.

Downloads

Download data is not yet available.

References

Hanenko, L., Storchak, K., Shlianchak, S., Vorokhob, M., & Pitaichuk, M. (2025). SLAM in navigation systems of autonomous mobile robots. Cybersecurity Providing in Information and Telecommunication Systems 2025, (3991), 173–182. https://ceur-ws.org/Vol-3991/

West, J., Maire, F., Browne, C., & Denman, S. (2020). Improved reinforcement learning with curriculum. Expert Systems with Applications, 158, 113515. https://doi.org/10.1016/j.eswa.2020.113515

Uppuluri, B., Patel, A., Mehta, N., Kamath, S., & Chakraborty, P. (2025). CuRLA: Curriculum learning based deep reinforcement learning for autonomous driving. In 17th International Conference on Agents and Artificial Intelligence (pp. 435–442). SCITEPRESS. https://doi.org/10.5220/0013147000003890

Freitag, K., Ceder, K., Laezza, R., Akesson, K., & Haghir Chehreghani, M. (2024). Sample-efficient curriculum reinforcement learning for complex reward functions. arXiv preprint. https://doi.org/10.48550/arXiv.2410.16790

Li, K., Lu, Y., & Meng, M. Q. H. (2021). Human-aware robot navigation via reinforcement learning with hindsight experience replay and curriculum learning. In 2021 IEEE International Conference on Robotics and Biomimetics (ROBIO). IEEE. https://doi.org/10.1109/robio54168.2021.9739519

Florensa, C., Held, D., Wulfmeier, M., Zhang, M., & Abbeel, P. (2017, October). Reverse curriculum generation for reinforcement learning. In Conference on Robot Learning (pp. 482–495). PMLR.

Zhu, K., & Zhang, T. (2021). Deep reinforcement learning based mobile robot navigation: A review. Tsinghua Science and Technology, 26(5), 674–691. https://doi.org/10.26599/tst.2021.9010012

Gao, J., Ye, W., Guo, J., & Li, Z. (2020). Deep reinforcement learning for indoor mobile robot path planning. Sensors, 20(19), 5493. https://doi.org/10.3390/s20195493

Hanenko, L. D., & Zhebka, V. V. (2024). Application of reinforcement learning methods for mobile robot path planning. Telecommunication and Information Technologies, 1, 16–25. https://doi.org/10.31673/2412-4338.2024.011625

Soviany, P., Ionescu, R. T., Rota, P., & Sebe, N. (2022). Curriculum learning: A survey. International Journal of Computer Vision. https://doi.org/10.1007/s11263-022-01611-x

Wang, X., Chen, Y., & Zhu, W. (2021). A survey on curriculum learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1. https://doi.org/10.1109/tpami.2021.3069908

Anca, M., Thomas, J. D., Pedamonti, D., Hansen, M., & Studley, M. (2023). Achieving goals using reward shaping and curriculum learning. In Lecture Notes in Networks and Systems (pp. 316–331). Springer Nature Switzerland. https://doi.org/10.1007/978-3-031-47454-5_24

Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O. (2017). Proximal policy optimization algorithms. arXiv preprint. https://doi.org/10.48550/arXiv.1707.06347

Hanenko, L., & Zhebka, V. (2025). Model of socially adaptive navigation of a mobile robot using reinforcement learning methods. Cybersecurity: Education, Science, Technology, 1(29). https://doi.org/10.28925/2663-4023.2025.29.907

Hanenko, L., & Zhebka, V. (2025). Development of a navigation system for an autonomous mobile robot using ROS 2. Cybersecurity: Education, Science, Technology, 4(28), 498–510. https://doi.org/10.28925/2663-4023.2025.28.824

Narvekar, S., Peng, B., Leonetti, M., Sinapov, J., Taylor, M. E., & Stone, P. (2020). Curriculum learning for reinforcement learning domains: A framework and survey. Journal of Machine Learning Research, 21(181), 1–50.

Downloads


Abstract views: 29

Published

2025-10-26

How to Cite

Hanenko, L., & Bushma, O. (2025). METHOD OF LEARNING OF AUTONOMOUS MOBILE ROBOTS BASED ON DRL AND CURRICULUM LEARNING. Electronic Professional Scientific Journal «Cybersecurity: Education, Science, Technique», 2(30), 568–582. https://doi.org/10.28925/2663-4023.2025.30.994