Coarse to fine-based image–point cloud fusion network for 3D object detectionShow others and affiliations
2024 (English)In: Information Fusion, ISSN 1566-2535, E-ISSN 1872-6305, Vol. 112, p. 1-12, article id 102551Article in journal (Refereed) Published
Abstract [en]
Enhancing original LiDAR point cloud features with virtual points has gained widespread attention in multimodal information fusion. However, existing methods struggle to leverage image depth information due to the sparse nature of point clouds, hindering proper alignment with camera-derived features. We propose a novel 3D object detection method that refines virtual point clouds using a coarse-to-fine approach, incorporating a dynamic 2D Gaussian distribution for better matching and a dynamic posterior density-aware RoI network for refined feature extraction. Our method achieves an average precision (AP) of 90.02% for moderate car detection on the KITTI validation set, outperforming state-of-the-art methods. Additionally, our approach yields AP scores of 86.58% and 82.16% for moderate and hard car detection categories on the KITTI test set, respectively. These results underscore the effectiveness of our method in addressing point cloud sparsity and enhancing 3D object detection performance. The code is available at https://github.com/ZhongkangZ/LidarIG. © 2024 Elsevier B.V.
Place, publisher, year, edition, pages
Amsterdam: Elsevier, 2024. Vol. 112, p. 1-12, article id 102551
Keywords [en]
Coarse to fine, Dynamic 2D Gaussian distribution, Image–point cloud fusion, Multimodal object detection, Quantized perception strategy
National Category
Electrical Engineering, Electronic Engineering, Information Engineering
Identifiers
URN: urn:nbn:se:hh:diva-54335DOI: 10.1016/j.inffus.2024.102551ISI: 001267740600001Scopus ID: 2-s2.0-85198004558OAI: oai:DiVA.org:hh-54335DiVA, id: diva2:1886153
Note
This work was supported by the Science and Technology Project of Hebei Education Department (No. QN2023014), Handan Science and Technology Research and Development Project (No. 3422304027), National Natural Science Foundation of China (No. 62373343), and Beijing Natural Science Foundation (No. L233036).
2024-07-302024-07-302024-07-30Bibliographically approved