中国民航大学学报 ›› 2024, Vol. 42 ›› Issue (5): 59-65.

• • 上一篇    下一篇

用于移动机器人路径规划的改进强化学习算法

张 威 1a,2,3 ,初泽源 1b ,杨玉涛 1a ,王 伟 1a
  

  1. 1. 中国民航大学 a. 航空工程学院;b. 安全科学与工程学院,天津 300300;2. 中国民航航空地面特种设备研究基地,
    天津
    300300;3. 民航智慧机场理论与系统重点实验室,广州
    510470
  • 收稿日期:2023-01-07 修回日期:2023-05-04 出版日期:2024-12-21 发布日期:2025-04-08
  • 作者简介:张威(1979— ),男,湖南衡阳人,教授,博士,研究方向为民航智能装备、机器人学等
  • 基金资助:
    国家自然科学基金民航联合研究基金重点项目(U2033208);天津市研究生科研创新项目(2021YJSS122)

Improved reinforcement learning algorithm for mobile robot path planning

ZHANG Wei 1a,2,3 , CHU Zeyuan1b , YANG Yutao1a , WANG Wei 1a
  

  1. 1a. College of Aeronautical Engineering; 1b. College of Safety Science and Engineering, CAUC, Tianjin 300300, China;
    2. Aviation Special Ground Equipment Research Base, CAAC,Tianjin 300300, China;
    3. Key Laboratory of Smart Airport Theory and System, CAAC, Guangzhou 510470, China
  • Received:2023-01-07 Revised:2023-05-04 Online:2024-12-21 Published:2025-04-08

摘要:

针对传统 Q-learning 算法规划出的路径存在平滑度差、收敛速度慢以及学习效率低的问题,本文提出一种用
于移动机器人路径规划的改进 Q-learning 算法。 首先,考虑障碍物密度及起始点相对位置来选择动作集,以
加快 Q-learning 算法的收敛速度;其次,为奖励函数加入一个连续的启发因子,启发因子由当前点与终点的
距离和当前点距地图中所有障碍物以及地图边界的距离组成;最后,在 Q 值表的初始化进程中引入尺度因子,
给移动机器人提供先验环境信息,并在栅格地图中对所提出的改进 Q-learning 算法进行仿真验证。 仿真结果
表明,改进 Q-learning 算法相比传统 Q-learning 算法收敛速度有明显提高,在复杂环境中的适应性更好,验
证了改进算法的优越性。

关键词:

Abstract:

Aiming at the problems of poor smoothness, slow convergence speed and low learning efficiency of the paths
planned by the traditional Q-learning algorithm, this paper proposes an improved Q-learning algorithm for mobile
robot path planning. Firstly, the density of obstacles and the relative position of the start point are considered to select the action set to accelerate the convergence speed of the Q-learning algorithm. Secondly, a continuous heuristic factor is added to the reward function, which consists of the distance between the current point and the end
point, and the distance of the current point from all the obstacles in the map as well as the boundary of the map. Finally, a scale factor is introduced into the initialization process of Q-value table to give the mobile robot with a priori environment information, and the proposed improved Q-learning algorithm is simulated and verified in a raster
map. The simulation results show that the convergence speed of the improved Q-learning algorithm is significantly
improved compared with the traditional Q-learning algorithm, and its adaptability in complex environments is better, which verifies the superiority of the improved algorithm.

Key words:

中图分类号: