1. MMAF-Net: Multi-view multi-stage adaptive fusion for multi-sensor 3D object detection.
- Author
-
Zhang, Wensheng, Shi, Hongli, Zhao, Yunche, Feng, Zhenan, and Lovreglio, Ruggiero
- Subjects
- *
OBJECT recognition (Computer vision) , *FEATURE extraction , *POINT cloud , *IMAGE fusion , *AUTONOMOUS vehicles , *MULTIMODAL user interfaces , *OBJECT tracking (Computer vision) , *MULTISENSOR data fusion - Abstract
In this paper, we propose a 3D object detection method called MMAF-Net that is based on the multi-view and multi-stage adaptive fusion of RGB images and LiDAR point cloud data. This is an end-to-end architecture, which combines the characteristics of RGB images, the front view of point clouds based on reflection intensity, and the bird's eye view of point clouds. It also adopts a multi-stage fusion approach of "data-level fusion + feature-level fusion" to fully exploit the strength of multimodal information. Our proposed method addresses key challenges found in current 3D object detection methods for autonomous driving, including insufficient feature extraction from multimodal data, rudimentary fusion techniques, and sensitivity to distance and occlusion. To ensure the comprehensive integration of multimodal information, we present a series of targeted fusion methods. Firstly, we propose a novel input form that encodes dense point cloud reflectivity information into the image to enhance its representational power. Secondly, we design the Region Attention Adaptive Fusion module utilizing an attention mechanism to guide the network in adaptively adjusting the importance of different features. Finally, we extend the 2D DIOU (Distance Intersection over Union) loss function to 3D and develop a joint regression loss based on 3D_DIOU and SmoothL1 to optimize the similarity between detected and ground truth boxes. The experimental results on the KITTI dataset demonstrate that MMAF-Net effectively addresses the challenges posed by highly obscured or crowded scenes while maintaining real-time performance and improving the detection accuracy of smaller and more difficult objects that are occluded at far distances. • A novel object detection method is proposed integrating RGB images and LiDAR data. • This study tests the method on the KITTI dataset. • This method tackles highly obscured or crowded scenes for autonomous driving. • This method improves the detection accuracy of smaller and occluded targets far away. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF