Point-Level Fusion and Channel Attention for 3D Object Detection in Autonomous Driving

自动驾驶中基于点级融合和通道注意力机制的三维目标检测

阅读:1

Abstract

As autonomous driving technology progresses, LiDAR-based 3D object detection has emerged as a fundamental element of environmental perception systems. PointPillars transforms point cloud data into a two-dimensional pseudo-image and employs a 2D CNN for efficient and precise detection. Nevertheless, this approach encounters two primary challenges: (1) the sparsity and disorganization of raw point clouds hinder the model's capacity to capture local features, thus impacting detection accuracy; and (2) existing models struggle to detect small objects within complex environments, particularly regarding orientation estimation. To address these issues, we propose two enhancements: (1) point-level fusion of LiDAR point clouds and RGB images, which integrates the semantic information of 2D images with the geometric features of 3D point clouds to improve model performance in intricate scenarios; (2) the incorporation of the Efficient Channel Attention mechanism to concentrate on essential features, particularly for small and sparse objects. Experimental results on the KITTI dataset indicate significant improvements, particularly in small object detection tasks, such as identifying pedestrians and cyclists. The enhanced model also demonstrates substantial gains in the Average Orientation Similarity (AOS) metric. These enhancements enhance the vehicle's ability to track and predict object trajectories in dynamic environments, critical for reliable recognition and decision-making.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。