Gait Learning for Hexapod Robot Facing Rough Terrain Based on Dueling-DQN Algorithm
DOI:
https://doi.org/10.62051/ijcsit.v2n1.44Keywords:
Gait Planning; Reinforcement Learning; Hexapod RobotAbstract
In the handling of dangerous goods in explosive environments, robots are increasingly being used instead of human operators. Robots designed for operation in explosive environments are mostly equipped with tracked structures, which, due to their limited terrain adaptability, struggle to movement rugged landscapes. Hexapod robots, with their excellent maneuverability and adaptability, possess advantages in completing hazardous material handling tasks in such rugged terrains. One current challenge lies in enabling hexapod robots to autonomously adjust their gaits to cope with rugged terrain. This paper proposes a gait learning method based on the Dueling Deep Q-Network (Dueling-DQN) algorithm to address the gait adjustment problem of hexapod robots in sloped, terraced, and rugged terrain. The method combines lidar data and reinforcement learning to extract features from the lidar data to determine terrain types and foot coordinates. Finally, the Dueling-DQN algorithm and redundant phase strategy are employed to facilitate the motion of hexapod robots in these three types of rugged terrain. Simulation and prototype experiments are conducted to evaluate the Dueling-DQN algorithm's performance in terms of rewards and stability margins for the three types of terrain. During algorithm training on sloped, terraced, and rugged terrain, stable rewards and positive stability margins are achieved after approximately 490 iterations. The effectiveness and feasibility of the proposed method are further validated through Gazebo simulation and prototype experiments. In the context of movement in rugged terrain within explosive environments, the gait learning method based on the Dueling-DQN algorithm offers valuable insights into the control of hexapod robots.
Downloads
References
Luccioni B M, Ambrosini R D, Danesi R F. Analysing explosive damage in an urban environment[J]. Proceedings of the Institution of Civil Engineers-Structures and Buildings, 2005, 158(1): 1-12.
Li Y, Li M, Zhu H, et al. Development and applications of rescue robots for explosion accidents in coal mines[J]. Journal of Field Robotics, 2020, 37(3): 466-489.
Liu K, Gao F, Chen Z, et al. Foothold Planning and Body Posture Adjustment Strategy of Hexapod Robot in Complex Terrain[C]//International Conference on Mechanism and Machine Science. Singapore: Springer Nature Singapore, 2022: 2023-2035.
Hou S, Cao C, Liu X, et al. Gait lateral network: Learning discriminative and compact representations for gait recognition[C]//European conference on computer vision. Cham: Springer International Publishing, 2020: 382-398.
Zu Xiaoqin, Chen Yaolu, Ran Dengyu. Autonomous gait learning for hexapod robot based on reward guidance [J]. Journal of Beijing University of Technology, 2021 (2).
Cizek P , Zoula M , Faigl J . Design, Construction, and Rough-Terrain Locomotion Control of Novel Hexapod Walking Robot with Four Degrees of Freedom per Leg[J]. IEEE Access, 2021, PP(99):1-1
Gao Y, Wang D, Wei W, et al. Constrained predictive tracking control for unmanned hexapod robot with tripod gait[J]. Drones, 2022, 6(9): 246.
Liu Y, Xu Y. Free gait planning of hexapod robot based on improved DQN algorithm[C]//2020 IEEE 2nd International Conference on Civil Aviation Safety and Information Technology (ICCASIT. IEEE, 2020: 488-491.
Youcef Z, Pierre C. Control of the trajectory of a hexapod robot based on distributed Q-learning[C]//2004 IEEE International Symposium on Industrial Electronics. IEEE, 2004, 1: 277-282.
Ouyang Y. Task offloading algorithm of vehicle edge computing environment based on Dueling-DQN[C]//Journal of Physics: Conference Series. IOP Publishing, 2021, 1873(1): 012046.
Gu Y, Zhu Z, Lv J, et al. DM-DQN: Dueling Munchausen deep Q network for robot path planning[J]. Complex & Intelligent Systems, 2023, 9(4): 4287-4300.
Peng B, Sun Q, Li S E, et al. End-to-end autonomous driving through dueling double deep Q-network[J]. Automotive Innovation, 2021, 4: 328-337.
Jiang W, Bao C, Xu G, et al. Research on autonomous obstacle avoidance and target tracking of UAV based on improved dueling DQN algorithm[C]//2021 China Automation Congress (CAC). IEEE, 2021: 5110-5115.
Wang Z, Schaul T, Hessel M, et al. Dueling network architectures for deep reinforcement learning[C]//International conference on machine learning. PMLR, 2016: 1995-2003.
Guo Jian, Liang Yongjie, Zhang Xiaojia, et al. Kinematics analysis and calculation of hexapod robot based on gait planning [J]. Machine Tools and Hydraulics, 2023, 51(05): 66-73.
Downloads
Published
Issue
Section
License
Copyright (c) 2024 Liuhongxu Chen, Ping Du, Pengfei Zhan, Bo Xie

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.







