Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection
Monocular 3D object detection is the most widely applied and challenging solution for autonomous driving, due to 2D images lacking 3D information. Existing methods are limited by inaccurate depth estimations by inequivalent supervised targets. The use of both depth and visual features also faces pro...
Saved in:
Main Authors: | , , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
PeerJ Inc.
2025-01-01
|
Series: | PeerJ Computer Science |
Subjects: | |
Online Access: | https://peerj.com/articles/cs-2656.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832577785675120640 |
---|---|
author | Zhijian Wang Jie Liu Yixiao Sun Xiang Zhou Boyan Sun Dehong Kong Jay Xu Xiaoping Yue Wenyu Zhang |
author_facet | Zhijian Wang Jie Liu Yixiao Sun Xiang Zhou Boyan Sun Dehong Kong Jay Xu Xiaoping Yue Wenyu Zhang |
author_sort | Zhijian Wang |
collection | DOAJ |
description | Monocular 3D object detection is the most widely applied and challenging solution for autonomous driving, due to 2D images lacking 3D information. Existing methods are limited by inaccurate depth estimations by inequivalent supervised targets. The use of both depth and visual features also faces problems of heterogeneous fusion. In this article, we propose Depth Detection Transformer (Depth-DETR), applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection. Depth-DETR introduces two additional depth encoders besides the visual encoder. Two depth encoders are supervised by ground truth depth and bounding box respectively, working independently to complement each other’s limitations and predicting more accurate target distances. Furthermore, Depth-DETR employs cross modal attention mechanisms to effectively fuse three different features. A parallel structure of two cross modal transformer is applied to fuse two depth features with visual features. Avoiding early fusion between two depth features enhances the final fused feature for better feature representations. Through multiple experimental validations, the Depth-DETR model has achieved highly competitive results in the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI) dataset, with an AP score of 17.49, representing its outstanding performance in 3D object detection. |
format | Article |
id | doaj-art-0479e626ee194c418c737d4897e0fe9e |
institution | Kabale University |
issn | 2376-5992 |
language | English |
publishDate | 2025-01-01 |
publisher | PeerJ Inc. |
record_format | Article |
series | PeerJ Computer Science |
spelling | doaj-art-0479e626ee194c418c737d4897e0fe9e2025-01-30T15:05:07ZengPeerJ Inc.PeerJ Computer Science2376-59922025-01-0111e265610.7717/peerj-cs.2656Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detectionZhijian Wang0Jie Liu1Yixiao Sun2Xiang Zhou3Boyan Sun4Dehong Kong5Jay Xu6Xiaoping Yue7Wenyu Zhang8School of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaAnshan Power Supply Company, Liaoning Electric Power Limited Company of State Grid, Anshan, Liaoning, ChinaSchool of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaInner Mongolia Electronic Information Vocational Technical College, Huhehaote, Neimenggu, ChinaSchool of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaSchool of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaSchool of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaSchool of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaSchool of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, Liaoning, ChinaMonocular 3D object detection is the most widely applied and challenging solution for autonomous driving, due to 2D images lacking 3D information. Existing methods are limited by inaccurate depth estimations by inequivalent supervised targets. The use of both depth and visual features also faces problems of heterogeneous fusion. In this article, we propose Depth Detection Transformer (Depth-DETR), applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection. Depth-DETR introduces two additional depth encoders besides the visual encoder. Two depth encoders are supervised by ground truth depth and bounding box respectively, working independently to complement each other’s limitations and predicting more accurate target distances. Furthermore, Depth-DETR employs cross modal attention mechanisms to effectively fuse three different features. A parallel structure of two cross modal transformer is applied to fuse two depth features with visual features. Avoiding early fusion between two depth features enhances the final fused feature for better feature representations. Through multiple experimental validations, the Depth-DETR model has achieved highly competitive results in the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI) dataset, with an AP score of 17.49, representing its outstanding performance in 3D object detection.https://peerj.com/articles/cs-2656.pdf3D object detectionTransformerDepth estimation |
spellingShingle | Zhijian Wang Jie Liu Yixiao Sun Xiang Zhou Boyan Sun Dehong Kong Jay Xu Xiaoping Yue Wenyu Zhang Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection PeerJ Computer Science 3D object detection Transformer Depth estimation |
title | Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection |
title_full | Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection |
title_fullStr | Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection |
title_full_unstemmed | Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection |
title_short | Applying auxiliary supervised depth-assisted transformer and cross modal attention fusion in monocular 3D object detection |
title_sort | applying auxiliary supervised depth assisted transformer and cross modal attention fusion in monocular 3d object detection |
topic | 3D object detection Transformer Depth estimation |
url | https://peerj.com/articles/cs-2656.pdf |
work_keys_str_mv | AT zhijianwang applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT jieliu applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT yixiaosun applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT xiangzhou applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT boyansun applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT dehongkong applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT jayxu applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT xiaopingyue applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection AT wenyuzhang applyingauxiliarysuperviseddepthassistedtransformerandcrossmodalattentionfusioninmonocular3dobjectdetection |