LLD-YOLO: A Low-Light Object Detection Algorithm Based on Dynamic Weighted Fusion of Shallow and Deep Features

Object detection in low-light scenarios has a wide range of applications, but existing algorithms often struggle to preserve the scarce low-level features in dark environments and exhibit limitations in localization accuracy for blurred edges and occluded objects, leading to suboptimal performance....

Full description

Saved in:
Bibliographic Details
Main Authors: Wenhao Cai, Yajun Chen, Xiaoyang Qiu, Meiqi Niu, Jianying Li
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10955203/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Object detection in low-light scenarios has a wide range of applications, but existing algorithms often struggle to preserve the scarce low-level features in dark environments and exhibit limitations in localization accuracy for blurred edges and occluded objects, leading to suboptimal performance. To address these challenges, we propose an improved neck structure, SRB-FPN, to achieve fine-grained cross-level semantic alignment and feature fusion, while also optimizing the regression loss function to develop LLD-YOLO, a detector specifically designed for low-light conditions. To enhance the representation of key feature units and dynamically optimize the fusion weights between shallow and deep features, we introduce the SDFBF module. To improve the diversity of receptive fields and strengthen the network’s multi-scale feature capture capability, we incorporate the DBB-C2f module. Furthermore, we integrate the hard-sample focusing property of Focaler IoU with the geometric perception advantages of MPDIoU, proposing Focal MPDIoU Loss to refine the localization of difficult samples and precisely capture bounding box variations. Ultimately, LLD-YOLO achieves an mAP50 of 70.0% on the ExDark dataset, outperforming the baseline by 2.7 percentage points. Extensive experiments on three public datasets, ExDark, NOD, and RTTS, further validate the superior performance of the proposed method in low-light conditions and its strong adaptability to foggy environments.
ISSN:2169-3536