Journal of System Simulation ›› 2025, Vol. 37 ›› Issue (3): 595-606.doi: 10.16182/j.issn1004731x.joss.24-0088
• Papers • Previous Articles Next Articles
Wang He1,2, Xu Jianing1, Yan Guangyu1,2
Received:2024-01-22
Revised:2024-03-07
Online:2025-03-17
Published:2025-03-21
Contact:
Yan Guangyu
CLC Number:
Wang He, Xu Jianing, Yan Guangyu. Research on Pedestrian Avoidance Strategy for AGV Based on Deep Reinforcement Learning[J]. Journal of System Simulation, 2025, 37(3): 595-606.
| 1 | Durrant-Whyte H, Bailey T. Simultaneous Localization and Mapping: Part I[J]. IEEE Robotics & Automation Magazine, 2006, 13(2): 99-110. |
| 2 | LaValle S M. Planning Algorithms[M]. Cambridge: Cambridge University Press, 2006. |
| 3 | Yi Guohong, Feng Zhili, Mei Tiancan, et al. Multi-AGVs Path Planning Based on Improved Ant Colony Algorithm[J]. The Journal of Supercomputing, 2019, 75(9): 5898-5913. |
| 4 | Fransen Karlijn, van Eekelena Joost. Efficient Path Planning for Automated Guided Vehicles Using A* (Astar) Algorithm Incorporating Turning Costs in Search Heuristic[J]. International Journal of Production Research, 2023, 61(3): 707-725. |
| 5 | Nwaonumah E, Samanta B. Deep Reinforcement Learning for Visual Navigation of Wheeled Mobile Robots[C]//2020 SoutheastCon. Piscataway: IEEE, 2020: 1-8. |
| 6 | Zhu Yuke, Mottaghi R, Kolve E, et al. Target-driven Visual Navigation in Indoor Scenes Using Deep Reinforcement Learning[C]//2017 IEEE International Conference on Robotics and Automation (ICRA). Piscataway: IEEE, 2017: 3357-3364. |
| 7 | Li Keyu, Xu Yangxin, Wang Jiankun, et al. SARL: Deep Reinforcement Learning Based Human-aware Navigation for Mobile Robot in Indoor Environments[C]//2019 IEEE International Conference on Robotics and Biomimetics (ROBIO). Piscataway: IEEE, 2019: 688-694. |
| 8 | Lu Xiaojun, Woo Hanwool, Faragasso Angela, et al. Socially Aware Robot Navigation in Crowds via Deep Reinforcement Learning with Resilient Reward Functions[J]. Advanced Robotics, 2022, 36(8): 388-403. |
| 9 | Wang S H, Wu Y H, Li T H S. Deep Reinforcement Learning with Pedestrian Trajectory Prediction Model for Service Robot Navigation in Crowded Environments[C]//2023 International Conference on Advanced Robotics and Intelligent Systems (ARIS). Piscataway: IEEE, 2023: 1-6. |
| 10 | Cimurs Reinis, Jin Han Lee, Il Hong Suh. Goal-oriented Obstacle Avoidance with Deep Reinforcement Learning in Continuous Action Space[J]. Electronics, 2020, 9(3): 411. |
| 11 | Lian Pengfei, Yuan Liang, Sun Lihui. Visual Navigation for Mobile Robots Based on Deep Reinforcement Learning[C]//2023 IEEE International Conference on Real-time Computing and Robotics (RCAR). Piscataway: IEEE, 2023: 650-656. |
| 12 | Xue Honghu, Hein Benedikt, Bakr Mohamed, et al. Using Deep Reinforcement Learning with Automatic Curriculum Learning for Mapless Navigation in Intralogistics[J]. Applied Sciences, 2022, 12(6): 3153. |
| 13 | Sunil Srivatsav Samsani, Mutahira Husna, Mannan Saeed Muhammad. Memory-based Crowd-aware Robot Navigation Using Deep Reinforcement Learning[J]. Complex & Intelligent Systems, 2023, 9(2): 2147-2158. |
| 14 | 林俊强, 王红军, 邹湘军, 等. 基于DPPO的移动采摘机器人避障路径规划及仿真[J]. 系统仿真学报, 2023, 35(8): 1692-1704. |
| Lin Junqiang, Wang Hongjun, Zou Xiangjun, et al. Obstacle Avoidance Path Planning and Simulation of Mobile Picking Robot Based on DPPO[J]. Journal of System Simulation, 2023, 35(8): 1692-1704. | |
| 15 | Schulman J, Wolski F, Dhariwal P, et al. Proximal Policy Optimization Algorithms[EB/OL]. (2017-08-28) [2023-12-20]. . |
| 16 | Schulman J, Levine S, Abbeel P, et al. Trust Region Policy Optimization[C]//Proceedings of the 32nd International Conference on Machine Learning. Chia Laguna Resort: PMLR, 2015: 1889-1897. |
| 17 | Mnih V, Adrià Puigdomènech Badia, Mirza M, et al. Asynchronous Methods for Deep Reinforcement Learning[C]//Proceedings of the 33rd International Conference on Machine Learning. Chia Laguna Resort: PMLR, 2016: 1928-1937. |
| 18 | 夏家伟, 朱旭芳, 罗亚松, 等. 基于深度强化学习的无人艇轨迹跟踪算法研究[J]. 华中科技大学学报(自然科学版), 2023, 51(5): 74-80. |
| Xia Jiawei, Zhu Xufang, Luo Yasong, et al. Study on Trajectory Tracking Algorithm of Unmanned Surface Vehicle Based on Deep Reinforcement Learning[J]. Journal of Huazhong University of Science and Technology(Natural Science Edition), 2023, 51(5): 74-80. | |
| 19 | Xiao Xin, Feng Xinlong. Multi-object Pedestrian Tracking Using Improved YOLOv8 and OC-SORT[J]. Sensors, 2023, 23(20): 8439. |
| 20 | Pacchierotti Elena, Christensen Henrik I, Jensfelt Patric. Evaluation of Passing Distance for Social Robots[C]//ROMAN 2006 - The 15th IEEE International Symposium on Robot and Human Interactive Communication. Piscataway: IEEE, 2006: 315-320. |
| 21 | 孙立香, 孙晓娴, 刘成菊, 等. 人群环境中基于深度强化学习的移动机器人避障算法[J]. 信息与控制, 2022, 51(1): 107-118. |
| Sun Lixiang, Sun Xiaoxian, Liu Chengju, et al. Obstacle Avoidance Algorithm for Mobile Robot Based on Deep Reinforcement Learning in Crowd Environment[J]. Information and Control, 2022, 51(1): 107-118. |
| [1] | Gong Feng, Jiang Tao, Zhang Qin, Liu Yu. Simulation and Optimization of Support Processes for Aircraft Fleet Launch Under Limited Resources [J]. Journal of System Simulation, 2025, 37(8): 1965-1977. |
| [2] | Lu Bin, Yang Xuan, Yang Zhenyu, Gao Xiaotian. Adaptive Sampling and Ghost Multi-scale Fusion for Lightweight Weld Defect Detection [J]. Journal of System Simulation, 2025, 37(8): 1978-1990. |
| [3] | Zhang Yue, Zhang Wenliang, Feng Qiang, Guo Xing, Ren Yi, Wang Zili. Combat-oriented Comprehensive Simulation and Verification Technology for Equipment System RMS [J]. Journal of System Simulation, 2025, 37(7): 1823-1835. |
| [4] | Wan Yuhang, Zhu Zilu, Zhong Chunfu, Liu Yongkui, Lin Tingyu, Zhang Lin. Dynamic Path Planning for Robotic Arms Based on an Improved PPO Algorithm [J]. Journal of System Simulation, 2025, 37(6): 1462-1473. |
| [5] | Zhang Yan, Li Binghua, Huo Tao, Liu Rong. Research on Robot Dynamic Obstacle Avoidance Method Based on Improved A* and Dynamic Window Algorithm [J]. Journal of System Simulation, 2025, 37(6): 1555-1564. |
| [6] | Wang Ziyi, Zhang Kai, Qian Dianwei, Liu Yuzhen. A DRL⁃based Approach for Distributed Equipment Nodes Selection [J]. Journal of System Simulation, 2025, 37(6): 1565-1573. |
| [7] | Wu Guohua, Zeng Jiaheng, Wang Dezhi, Zheng Long, Zou Wei. A Quadrotor Trajectory Tracking Control Method Based on Deep Reinforcement Learning [J]. Journal of System Simulation, 2025, 37(5): 1169-1187. |
| [8] | Li Jie, Liu Yang, Li Liang, Su Bengan, Wei Jialong, Zhou Guangda, Shi Yanmin, Zhao Zhen. Remote Sensing Small Object Detection Based on Cross-stage Two-branch Feature Aggregation [J]. Journal of System Simulation, 2025, 37(4): 1025-1040. |
| [9] | Zhang Sen, Dai Qiangqiang. UAV Path Planning Based on Improved Deep Deterministic Policy Gradients [J]. Journal of System Simulation, 2025, 37(4): 875-881. |
| [10] | Li Min, Zhang Sen, Zeng Xiangguang, Wang Gang, Zhang Tongwei, Xie Dijie, Ren Wenzhe, Zhang Tao. Trajectory Planning of Quadruped Robot Over Obstacle with Single Leg Based on Deep Reinforcement Learning [J]. Journal of System Simulation, 2025, 37(4): 895-909. |
| [11] | Zhang Bin, Lei Yonglin, Li Qun, Gao Yuan, Chen Yong, Zhu Jiajun, Bao Chenlong. Reinforcement Learning Modeling of Missile Penetration Decision Based on Combat Simulation [J]. Journal of System Simulation, 2025, 37(3): 763-774. |
| [12] | Huang Sijin, Wen Jia, Chen Zheyi. Intelligent Service Migration towards MEC-based IoV Systems [J]. Journal of System Simulation, 2025, 37(2): 379-391. |
| [13] | Fei Shuaidi, Cai Changlong, Liu Fei, Chen Minghui, Liu Xiaoming. Research on the Target Allocation Method for Air Defense and Anti-missile Defense of Naval Ships [J]. Journal of System Simulation, 2025, 37(2): 508-516. |
| [14] | Peng Yi, Lei Yunkui, Yang Qingqing, Li Hui, Wang Jianming. Improved PID Search Algorithm for UAV Path Planning in Mountainous Environments [J]. Journal of System Simulation, 2025, 37(12): 3075-3086. |
| [15] | Wu Shuheng, Liu Yongkui, Zhang Lin, Xiao Yingying, Wang Lihui. Lightweight Assembly Workpiece Detection Algorithm Based on Improved YOLOv8 [J]. Journal of System Simulation, 2025, 37(12): 3099-3111. |
| Viewed | ||||||
|
Full text |
|
|||||
|
Abstract |
|
|||||