Gait Learning for Hexapod Robot Facing Rough Terrain Based on Dueling-DQN Algorithm

Authors

  • Liuhongxu Chen
  • Ping Du
  • Pengfei Zhan
  • Bo Xie

DOI:

https://doi.org/10.62051/ijcsit.v2n1.44

Keywords:

Gait Planning; Reinforcement Learning; Hexapod Robot

Abstract

In the handling of dangerous goods in explosive environments, robots are increasingly being used instead of human operators. Robots designed for operation in explosive environments are mostly equipped with tracked structures, which, due to their limited terrain adaptability, struggle to movement rugged landscapes. Hexapod robots, with their excellent maneuverability and adaptability, possess advantages in completing hazardous material handling tasks in such rugged terrains. One current challenge lies in enabling hexapod robots to autonomously adjust their gaits to cope with rugged terrain. This paper proposes a gait learning method based on the Dueling Deep Q-Network (Dueling-DQN) algorithm to address the gait adjustment problem of hexapod robots in sloped, terraced, and rugged terrain. The method combines lidar data and reinforcement learning to extract features from the lidar data to determine terrain types and foot coordinates. Finally, the Dueling-DQN algorithm and redundant phase strategy are employed to facilitate the motion of hexapod robots in these three types of rugged terrain. Simulation and prototype experiments are conducted to evaluate the Dueling-DQN algorithm's performance in terms of rewards and stability margins for the three types of terrain. During algorithm training on sloped, terraced, and rugged terrain, stable rewards and positive stability margins are achieved after approximately 490 iterations. The effectiveness and feasibility of the proposed method are further validated through Gazebo simulation and prototype experiments. In the context of movement in rugged terrain within explosive environments, the gait learning method based on the Dueling-DQN algorithm offers valuable insights into the control of hexapod robots.

Downloads

Download data is not yet available.

References

Luccioni B M, Ambrosini R D, Danesi R F. Analysing explosive damage in an urban environment[J]. Proceedings of the Institution of Civil Engineers-Structures and Buildings, 2005, 158(1): 1-12.

Li Y, Li M, Zhu H, et al. Development and applications of rescue robots for explosion accidents in coal mines[J]. Journal of Field Robotics, 2020, 37(3): 466-489.

Liu K, Gao F, Chen Z, et al. Foothold Planning and Body Posture Adjustment Strategy of Hexapod Robot in Complex Terrain[C]//International Conference on Mechanism and Machine Science. Singapore: Springer Nature Singapore, 2022: 2023-2035.

Hou S, Cao C, Liu X, et al. Gait lateral network: Learning discriminative and compact representations for gait recognition[C]//European conference on computer vision. Cham: Springer International Publishing, 2020: 382-398.

Zu Xiaoqin, Chen Yaolu, Ran Dengyu. Autonomous gait learning for hexapod robot based on reward guidance [J]. Journal of Beijing University of Technology, 2021 (2).

Cizek P , Zoula M , Faigl J . Design, Construction, and Rough-Terrain Locomotion Control of Novel Hexapod Walking Robot with Four Degrees of Freedom per Leg[J]. IEEE Access, 2021, PP(99):1-1

Gao Y, Wang D, Wei W, et al. Constrained predictive tracking control for unmanned hexapod robot with tripod gait[J]. Drones, 2022, 6(9): 246.

Liu Y, Xu Y. Free gait planning of hexapod robot based on improved DQN algorithm[C]//2020 IEEE 2nd International Conference on Civil Aviation Safety and Information Technology (ICCASIT. IEEE, 2020: 488-491.

Youcef Z, Pierre C. Control of the trajectory of a hexapod robot based on distributed Q-learning[C]//2004 IEEE International Symposium on Industrial Electronics. IEEE, 2004, 1: 277-282.

Ouyang Y. Task offloading algorithm of vehicle edge computing environment based on Dueling-DQN[C]//Journal of Physics: Conference Series. IOP Publishing, 2021, 1873(1): 012046.

Gu Y, Zhu Z, Lv J, et al. DM-DQN: Dueling Munchausen deep Q network for robot path planning[J]. Complex & Intelligent Systems, 2023, 9(4): 4287-4300.

Peng B, Sun Q, Li S E, et al. End-to-end autonomous driving through dueling double deep Q-network[J]. Automotive Innovation, 2021, 4: 328-337.

Jiang W, Bao C, Xu G, et al. Research on autonomous obstacle avoidance and target tracking of UAV based on improved dueling DQN algorithm[C]//2021 China Automation Congress (CAC). IEEE, 2021: 5110-5115.

Wang Z, Schaul T, Hessel M, et al. Dueling network architectures for deep reinforcement learning[C]//International conference on machine learning. PMLR, 2016: 1995-2003.

Guo Jian, Liang Yongjie, Zhang Xiaojia, et al. Kinematics analysis and calculation of hexapod robot based on gait planning [J]. Machine Tools and Hydraulics, 2023, 51(05): 66-73.

Downloads

Published

25-03-2024

Issue

Section

Articles

How to Cite

Chen, L., Du, P., Zhan, P., & Xie, B. (2024). Gait Learning for Hexapod Robot Facing Rough Terrain Based on Dueling-DQN Algorithm. International Journal of Computer Science and Information Technology, 2(1), 408-424. https://doi.org/10.62051/ijcsit.v2n1.44