Research on Multi-modal Point Cloud Completion Task

Authors

  • Wentian Chen

DOI:

https://doi.org/10.62051/ijcsit.v3n2.44

Keywords:

Multi-modal fusion, Point cloud completion, Deep learning, Self-attention mechanism, Generative adversarial networks

Abstract

With the wide growth of 3D data applications, point cloud completion is particularly critical in the fields of autonomous driving and robot navigation. Aiming at the problem that point cloud data is easily affected by occlusion and noise, these papers propose is a completion method based on multi-modal fusion strategy, which combines 3D lidar and structured light scanning modality information to achieve more accurate point cloud completion. Based on the point cloud data, we deeply analyze the sparsity and unstructured characteristics of the point cloud, explore the progress of multi-modal representation learning, and effectively apply GAN and self-attention mechanism to the completion task. This paper designs a hybrid encoder-decoder network architecture and integrates a special multi-modal feature extraction module, which can effectively capture the supplementary information from different modalities and improve the feature expression ability with the help of the self-attention mechanism. The experimental results show that the proposed multi-modal point cloud completion method has better completion effects than the current SOTA model, especially in dealing with point cloud data with highly missing and complex scenes.

Downloads

Download data is not yet available.

References

Hang Lingxiao. Research on indoor scene parsing Algorithm Based on RGB-D multi-modal Images. 2019.

Fangneng Zhan,Yingchen Yu, Rongliang Wu, et al. Multimodal Image Synthesis and Editing: The Generative AI Era [D]. 2023.

Xu Guanghui. Research on complex Scene Reasoning Algorithm Based on multi-modal attention mechanism. 2021.

Song Wenlin. Research on 3D laser point cloud face recognition fused with multi-modal features [J]. Laser Journal, 2023. (in Chinese)

H Zhang, M Zhang, C Li, et al. Research on Blended Teaching Based on"MOOC+SPOC+multimodal classroom" [D]. 2021

MENG Yue, Li Shixin, Chen Fankai, et al. Review of Multimodal Fusion of Point cloud and Image in Autonomous Driving [J]. Computer Science and Application, 2023.

Fu Si Meng. Research on multimodal medical image registration and fusion in neurosurgery [J]. 2020.

Wang Q., Hu X., Lei S., et al. Research on downhole visible light modeling and localization based on point cloud technology [D]. 2024.

Li Xiaojun. Research on multi-modal medical image fusion algorithm based on NSST [J]. 2020.

Li Qiming, REN Jieji, PEI Xiaohan, et al. High-precision point cloud registration algorithm for weak texture surface based on multi-modal data collaboration [J]. Acta Optica Sinica, 2022.

WEI Yun-peng. Research and application of ensemble Learners Based on Multi-modal evolutionary Computation.,2019.

Guo Xu, Mai Ridan Wu Shouer, Guranbaier Turhong. Research on sentiment analysis Algorithm Based on multi-modal fusion [J]. Computer Engineering and Application, 2024. (in Chinese)

Chen Fan. Research on 3D object recognition and 6D Pose Estimation Based on depth and texture information fusion. 2020.

LIU Liqiang. Research on visual 3D Reconstruction based on Deep Learning [J]. 2019.

Xie L J. Research on robust building modeling method based on image dense matching point cloud [J]. 2022.

Downloads

Published

19-07-2024

Issue

Section

Articles

How to Cite

Chen, W. (2024). Research on Multi-modal Point Cloud Completion Task. International Journal of Computer Science and Information Technology, 3(2), 402-411. https://doi.org/10.62051/ijcsit.v3n2.44