SDPSNet: An Efficient 3D Object Detection Based on Spatial Dynamic Pruning Sparse Convolution and Self-Attention Feature Diffusion

In the field of autonomous driving, the technology of 3D object detection using LiDAR point clouds has been widely implemented. However, existing detectors are faced with challenges due to redundant overhead when processing large amounts of background points in the depth perception field, and insuff...

Full description

Saved in:
Bibliographic Details
Main Authors: Meng Wang, Qianlei Yu, Haipeng Liu
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10962225/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In the field of autonomous driving, the technology of 3D object detection using LiDAR point clouds has been widely implemented. However, existing detectors are faced with challenges due to redundant overhead when processing large amounts of background points in the depth perception field, and insufficient feature mapping in the central regions of large objects. To address these issues, this paper proposes a novel 3D object detection model, SDPSNet, which combines spatial dynamic pruning and self-attention feature diffusion to reduce data redundancy and improve the representation of central features. SDPSNet introduces a spatial dynamic pruning sparse convolution strategy to optimise the inference efficiency and data redundancy of the detector. Meanwhile, the self-attention feature diffusion module is used to effectively diffuse the features of the object edges to the central region, providing a richer and more accurate representation for modelling the centre of large objects. Extensive experimental results have been conducted on the Waymo Open, nuScenes and Argoverse2 datasets. On Argoverse2, SDPSNet achieves a 2.8% increase in mAP over the previous hybrid detector, HEDNet, while being <inline-formula> <tex-math notation="LaTeX">$2.4\times $ </tex-math></inline-formula> faster. It outperforms the previously popular sparse detector FSDv2 by 2.3% in mAP and is <inline-formula> <tex-math notation="LaTeX">$1.5\times $ </tex-math></inline-formula> faster. It also shows a 0.2% improvement in mAP and is <inline-formula> <tex-math notation="LaTeX">$1.2\times $ </tex-math></inline-formula> faster than the current best-performing sparse detector, SAFDNet.
ISSN:2169-3536