Affiliation:
1. School of Electronic, Electrical Engineer and Physics, Fujian University of Technology, Fuzhou 350118, China
2. Fujian Province Industrial Integrated Automation Industry Technology Development Base, Fuzhou 350118, China
Abstract
Path planning and obstacle avoidance are fundamental problems in unmanned ground vehicle path planning. Aiming at the limitations of Deep Reinforcement Learning (DRL) algorithms in unmanned ground vehicle path planning, such as low sampling rate, insufficient exploration, and unstable training, this paper proposes an improved algorithm called Dual Priority Experience and Ornstein–Uhlenbeck Soft Actor-Critic (DPEOU-SAC) based on Ornstein–Uhlenbeck (OU noise) and double-factor prioritized sampling experience replay (DPE) with the introduction of expert experience, which is used to help the agent achieve faster and better path planning and obstacle avoidance. Firstly, OU noise enhances the agent’s action selection quality through temporal correlation, thereby improving the agent’s detection performance in complex unknown environments. Meanwhile, the experience replay is based on double-factor preferential sampling, which has better sample continuity and sample utilization. Then, the introduced expert experience can help the agent to find the optimal path with faster training speed and avoid falling into a local optimum, thus achieving stable training. Finally, the proposed DPEOU-SAC algorithm is tested against other deep reinforcement learning algorithms in four different simulation environments. The experimental results show that the convergence speed of DPEOU-SAC is 88.99% higher than the traditional SAC algorithm, and the shortest path length of DPEOU-SAC is 27.24, which is shorter than that of SAC.
Funder
Fujian University Industry–University-Research Joint Innovation Project
Fujian University Industry–University Cooperation Science and Technology Program
Reference31 articles.
1. Wu, M., Zhu, S., Li, C., Zhu, J., Chen, Y., Liu, X., and Liu, R. (2024). UAV-Mounted RIS-Aided Mobile Edge Computing System: A DDQN-Based Optimization Approach. Drones, 8, Available online: https://www.mdpi.com/2504-446X/8/5/184.
2. Multi-robot formation control and object transport in dynamic environments via constrained optimization;Baker;Int. J. Robot. Res.,2017
3. Modeling and solving the repair crew scheduling for the damaged road networks based on Q-learning;Su;Acta Autom. Sin.,2020
4. Research progress on human-robot skill transfer;Zeng;Acta Autom. Sin.,2019
5. Gasparetto, A., Boscariol, P., Lanzutti, A., and Vidoni, R. (2015). Path planning and trajectory planning algorithms: A general overview. Motion and Operation Planning of Robotic Systems: Background and Practical Approaches, Springer.