Citation: | LYU Youhao, JIA Yuanjun, ZHUANG Yuan, DONG Qi. Obstacle avoidance approach for quadruped robot based on multi-modal information fusion[J]. Chinese Journal of Engineering, 2024, 46(8): 1426-1433. DOI: 10.13374/j.issn2095-9389.2023.07.01.002 |
This paper proposes a multimodal information fusion neural network model that integrates visual, radar, and proprioceptive information. The model uses a spatial crossmodal attention mechanism to fuse the information, allowing the robot to focus on the most relevant information for obstacle avoidance. The attention mechanism enables the robot to selectively focus on the most relevant informative sensory inputs, which improves its ability to navigate complex terrain. The proposed method was evaluated using multiple experiments in challenging simulated environments, and the results showed a significant improvement in the obstacle avoidance success rate. The proposed method uses an actor–critic architecture and a proximal policy optimization (PPO) algorithm to train the robot in a simulated environment. The training process aims to reduce the difference between the robot’s performance in simulated and real-world environments. To achieve this, we randomly adjust the simulation environment’s parameters and add random noise to the robot’s sensory inputs. This approach allows the robot to learn a robust planning strategy that can be deployed in real-world environments. The multimodal information fusion neural network model is designed using a transformer-based architecture. The model shares the encoding of three types of tokens and generates features for the robot’s proprioceptive, visual, and point cloud inputs. The transformer encoder layers are stacked such that the token information from the three modalities can be fuzed at multiple levels. To balance the information from the three modalities, we first separately collect information for each modality and calculate the average value of all tokens from the same modality to obtain a single feature vector. This multimodal information fusion approach improves the robot’s decision-making capabilities in complex environments. The novelty of the proposed method lies in the introduction of a spatial crossmodal attention mechanism that allows the robot to selectively attend to the most informative sensory inputs. This attention mechanism improves the robot’s ability to navigate complex terrain and provides a certain degree of reliability for the quadruped robot in dynamic unknown environments. The combination of multimodal information fusion and attention mechanism enables the robot to adapt better to complex environments, thus improving its obstacle avoidance capabilities. Therefore, the proposed method provides a promising approach for improving the obstacle avoidance capabilities of quadruped robots in complex environments. The proposed method is based on the multimodal information fusion neural network model and spatial crossmodal attention mechanism. The experimental results demonstrate the effectiveness of the proposed method in improving the robot’s obstacle avoidance success rate. Moreover, the potential applications of the proposed method include search and rescue missions, exploration, and surveillance in complex environments.
[1] |
Miki T, Lee J, Hwangbo J, et al. Learning robust perceptive locomotion for quadrupedal robots in the wild. Sci Robot, 2022, 7(62): eabk2822 doi: 10.1126/scirobotics.abk2822
|
[2] |
Grandia R, Jenelten F, Yang S H, et al. Perceptive locomotion through nonlinear model-predictive control. IEEE Trans Robotics, 2023, 39(5): 3402 doi: 10.1109/TRO.2023.3275384
|
[3] |
Zhou S Y, Xi J Q, McDaniel M W, et al. Self-supervised learning to visually detect terrain surfaces for autonomous robots operating in forested terrain. J Field Robotics, 2012, 29(2): 277 doi: 10.1002/rob.21417
|
[4] |
Velasquez A E B, Higuti V A H, Gasparino M V, et al. Multi-sensor fusion based robust row following for compact agricultural robots [J/OL]. arXiv preprint (2021-01-28) [2023-07-01]. https://arxiv.org/abs/2106.15029
|
[5] |
Ji T C, Sivakumar A N, Chowdhary G, et al. Proactive anomaly detection for robot navigation with multi-sensor fusion. IEEE Robotics Autom Lett, 2022, 7(2): 4975 doi: 10.1109/LRA.2022.3153989
|
[6] |
Ji T C, Vuppala S T, Chowdhary G, et al. Multi-modal anomaly detection for unstructured and uncertain environments [J/OL]. arXiv: prepint (2020-12-15) [2023-07-01]. https://arxiv.org/abs/2012.08637
|
[7] |
Weerakoon K, Sathyamoorthy A J, Patel U, et al. TERP: Reliable planning in uneven outdoor environments using deep reinforcement learning // 2022 International Conference on Robotics and Automation (ICRA). Philadelphia, 2022: 9447
|
[8] |
Kahn G, Abbeel P, Levine S. BADGR: An autonomous self-supervised learning-based navigation system. IEEE Robotics Autom Lett, 2021, 6(2): 1312 doi: 10.1109/LRA.2021.3057023
|
[9] |
Kahn G, Abbeel P, Levine S. LaND: Learning to navigate from disengagements. IEEE Robotics Autom Lett, 2021, 6(2): 1872 doi: 10.1109/LRA.2021.3060404
|
[10] |
Bledt G, Kim S. Extracting legged locomotion heuristics with regularized predictive control // 2020 IEEE International Conference on Robotics and Automation (ICRA). Paris, 2020: 406
|
[11] |
Carius J, Ranftl R, Koltun V, et al. Trajectory optimization for legged robots with slipping motions. IEEE Robotics Autom Lett, 2019, 4(3): 3013 doi: 10.1109/LRA.2019.2923967
|
[12] |
Li T Y, Calandra R, Pathak D, et al. Planning in learned latent action spaces for generalizable legged locomotion. IEEE Robotics Autom Lett, 2021, 6(2): 2682 doi: 10.1109/LRA.2021.3062342
|
[13] |
Lu G L, Chen T, Rong X W, et al. Whole-body motion planning and control of a quadruped robot for challenging terrain. J Field Robotics, 2023, 40(6): 1657 doi: 10.1002/rob.22197
|
[14] |
Jiang X Y, Chi W C, Zheng Y, et al. Locomotion generation for quadruped robots on challenging terrains via quadratic programming. Auton Robot, 2023, 47(1): 51 doi: 10.1007/s10514-022-10068-3
|
[15] |
Xu S H, Zhu L J, Ho C P. Learning efficient and robust multi-modal quadruped locomotion: A hierarchical approach // 2022 International Conference on Robotics and Automation (ICRA). Philadelphia, 2022: 4649
|
[16] |
Aladem M, Baek S, Rawashdeh S A. Evaluation of image enhancement techniques for vision-based navigation under low illumination. J Robotics, 2019, 2019: 1
|
[17] |
Song H L, Li A, Wang T, et al. Multimodal deep reinforcement learning with auxiliary task for obstacle avoidance of indoor mobile robot. Sensors, 2021, 21(4): 1363 doi: 10.3390/s21041363
|
[18] |
Guan T R, Kothandaraman D, Chandra R, et al. GA-nav: Efficient terrain segmentation for robot navigation in unstructured outdoor environments. IEEE Robotics Autom Lett, 2022, 7(3): 8138 doi: 10.1109/LRA.2022.3187278
|
[19] |
Mansouri S S, Kanellakis C, Kominiak D, et al. Deploying MAVs for autonomous navigation in dark underground mine environments. Robotics Auton Syst, 2020, 126: 103472 doi: 10.1016/j.robot.2020.103472
|
[20] |
Cai P D, Wang S K, Sun Y X, et al. Probabilistic end-to-end vehicle navigation in complex dynamic environments with multimodal sensor fusion. IEEE Robotics Autom Lett, 2020, 5(3): 4218
|
[21] |
Qu Y H, Yang M H, Zhang J Q, et al. An outline of multi-sensor fusion methods for mobile agents indoor navigation. Sensors, 2021, 21(5): 1605 doi: 10.3390/s21051605
|
[22] |
He K M, Zhang X Y, Ren S Q, et al. Deep residual learning for image recognition // 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, 2016: 770
|
[23] |
Qi C R, Su H, Mo K C, et al. Pointnet: deep learning on point sets for 3D classification and segmentation // 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, 2017: 652
|
[24] |
Huang X Q, Deng H, Zhang W, et al. Towards multi-modal perception-based navigation: A deep reinforcement learning method. IEEE Robotics Autom Lett, 2021, 6(3): 4986 doi: 10.1109/LRA.2021.3064461
|
[25] |
Park D, Kim H, Kemp C C. Multimodal anomaly detection for assistive robots. Auton Robot, 2019, 43: 611 doi: 10.1007/s10514-018-9733-6
|
[26] |
Wellhausen L, Ranftl R, Hutter M. Safe robot navigation via multi-modal anomaly detection. IEEE Robotics Autom Lett, 2020, 5(2): 1326 doi: 10.1109/LRA.2020.2967706
|
[27] |
Sorokin M, Tan J, Liu C K, et al. Learning to navigate sidewalks in outdoor environments. IEEE Robotics Autom Lett, 2022, 7(2): 3906 doi: 10.1109/LRA.2022.3145947
|
[28] |
Rusli L, Nurhalim B, Rusyadi R. Vision-based vanishing point detection of autonomous navigation of mobile robot for outdoor applications. J Mechatron Electr Power Veh Technol, 2021, 12(2): 117 doi: 10.14203/j.mev.2021.v12.117-125
|
[29] |
Li Z, Li B, Liang Q X, et al. A quadruped robot obstacle avoidance and personnel following strategy based on ultra-wideband and three-dimensional laser radar. Int J Adv Robotic Syst, 2022, 19(4): 17298806221114705
|
1. |
王颖. 基于机器视觉的棉花智能采摘机器人技术概述. 中国棉花. 2025(01): 52-55 .
![]() | |
2. |
王振保,王子健,徐新喜,刘鑫,程韬,刘培朋,赵秀国,苏琛. 急救机器人关键技术及装备发展现状. 医疗卫生装备. 2025(03): 96-114 .
![]() |