期刊论文详细信息
Frontiers in Neurorobotics
Deep reinforcement learning-aided autonomous navigation with landmark generators
Neuroscience
Xuanzhi Wang1  Yankang Sun1  Yuyang Xie1  Jiang Bin1  Jian Xiao2 
[1] Department of Integrated Circuit Science and Engineering, Nanjing University of Posts and Telecommunications, Nanjing, China;null;
关键词: deep reinforcement learning (DRL);    mobile robots;    trajectory planning;    autonomous navigation;    simultaneous localization and mapping;   
DOI  :  10.3389/fnbot.2023.1200214
 received in 2023-04-04, accepted in 2023-08-08,  发布年份 2023
来源: Frontiers
PDF
【 摘 要 】

Mobile robots are playing an increasingly significant role in social life and industrial production, such as searching and rescuing robots, autonomous exploration of sweeping robots, and so on. Improving the accuracy of autonomous navigation of mobile robots is a hot issue to be solved. However, traditional navigation methods are unable to realize crash-free navigation in an environment with dynamic obstacles, more and more scholars are gradually using autonomous navigation based on deep reinforcement learning (DRL) to replace overly conservative traditional methods. But on the other hand, DRL's training time is too long, and the lack of long-term memory easily leads the robot to a dead end, which makes its application in the actual scene more difficult. To shorten training time and prevent mobile robots from getting stuck and spinning around, we design a new robot autonomous navigation framework which combines the traditional global planning and the local planning based on DRL. Therefore, the entire navigation process can be transformed into first using traditional navigation algorithms to find the global path, then searching for several high-value landmarks on the global path, and then using the DRL algorithm to move the mobile robot toward the designated landmarks to complete the final navigation, which makes the robot training difficulty greatly reduced. Furthermore, in order to improve the lack of long-term memory in deep reinforcement learning, we design a feature extraction network containing memory modules to preserve the long-term dependence of input features. Through comparing our methods with traditional navigation methods and reinforcement learning based on end-to-end depth navigation methods, it shows that while the number of dynamic obstacles is large and obstacles are rapidly moving, our proposed method is, on average, 20% better than the second ranked method in navigation efficiency (navigation time and navigation paths' length), 34% better than the second ranked method in safety (collision times), 26.6% higher than the second ranked method in success rate, and shows strong robustness.

【 授权许可】

Unknown   
Copyright © 2023 Wang, Sun, Xie, Bin and Xiao.

【 预 览 】
附件列表
Files Size Format View
RO202310105116967ZK.pdf 2782KB PDF download
  文献评价指标  
  下载次数:10次 浏览次数:0次