Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
This study proposed a vision-based underwater vertical docking guidance and positioning method to address docking control challenges for human-operated vehicles (HOVs) and unmanned underwater vehicles (UUVs) under complex underwater visual conditions. A cascaded detection and positioning strategy in...
Saved in:
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2025-01-01
|
Series: | Journal of Marine Science and Engineering |
Subjects: | |
Online Access: | https://www.mdpi.com/2077-1312/13/1/102 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832588232299118592 |
---|---|
author | Tian Ni Can Sima Wenzhong Zhang Junlin Wang Jia Guo Lindan Zhang |
author_facet | Tian Ni Can Sima Wenzhong Zhang Junlin Wang Jia Guo Lindan Zhang |
author_sort | Tian Ni |
collection | DOAJ |
description | This study proposed a vision-based underwater vertical docking guidance and positioning method to address docking control challenges for human-operated vehicles (HOVs) and unmanned underwater vehicles (UUVs) under complex underwater visual conditions. A cascaded detection and positioning strategy incorporating fused active and passive markers enabled real-time detection of the relative position and pose between the UUV and docking station (DS). A novel deep learning-based network model, YOLO-D, was developed to detect docking markers in real time. YOLO-D employed the Adaptive Kernel Convolution Module (AKConv) to dynamically adjust the sample shapes and sizes and optimize the target feature detection across various scales and regions. It integrated the Context Aggregation Network (CONTAINER) to enhance small-target detection and overall image accuracy, while the bidirectional feature pyramid network (BiFPN) facilitated effective cross-scale feature fusion, improving detection precision for multi-scale and fuzzy targets. In addition, an underwater docking positioning algorithm leveraging multiple markers was implemented. Tests on an underwater docking markers dataset demonstrated that YOLO-D achieved a detection accuracy of mAP@0.5 to 94.5%, surpassing the baseline YOLOv11n with improvements of 1.5% in precision, 5% in recall, and 4.2% in mAP@0.5. Pool experiments verified the feasibility of the method, achieving a 90% success rate for single-attempt docking and recovery. The proposed approach offered an accurate and efficient solution for underwater docking guidance and target detection, which is of great significance for improving the safety of docking. |
format | Article |
id | doaj-art-c6f48eb94638434aac54800ce18b4b6f |
institution | Kabale University |
issn | 2077-1312 |
language | English |
publishDate | 2025-01-01 |
publisher | MDPI AG |
record_format | Article |
series | Journal of Marine Science and Engineering |
spelling | doaj-art-c6f48eb94638434aac54800ce18b4b6f2025-01-24T13:36:51ZengMDPI AGJournal of Marine Science and Engineering2077-13122025-01-0113110210.3390/jmse13010102Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-DTian Ni0Can Sima1Wenzhong Zhang2Junlin Wang3Jia Guo4Lindan Zhang5China Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaThis study proposed a vision-based underwater vertical docking guidance and positioning method to address docking control challenges for human-operated vehicles (HOVs) and unmanned underwater vehicles (UUVs) under complex underwater visual conditions. A cascaded detection and positioning strategy incorporating fused active and passive markers enabled real-time detection of the relative position and pose between the UUV and docking station (DS). A novel deep learning-based network model, YOLO-D, was developed to detect docking markers in real time. YOLO-D employed the Adaptive Kernel Convolution Module (AKConv) to dynamically adjust the sample shapes and sizes and optimize the target feature detection across various scales and regions. It integrated the Context Aggregation Network (CONTAINER) to enhance small-target detection and overall image accuracy, while the bidirectional feature pyramid network (BiFPN) facilitated effective cross-scale feature fusion, improving detection precision for multi-scale and fuzzy targets. In addition, an underwater docking positioning algorithm leveraging multiple markers was implemented. Tests on an underwater docking markers dataset demonstrated that YOLO-D achieved a detection accuracy of mAP@0.5 to 94.5%, surpassing the baseline YOLOv11n with improvements of 1.5% in precision, 5% in recall, and 4.2% in mAP@0.5. Pool experiments verified the feasibility of the method, achieving a 90% success rate for single-attempt docking and recovery. The proposed approach offered an accurate and efficient solution for underwater docking guidance and target detection, which is of great significance for improving the safety of docking.https://www.mdpi.com/2077-1312/13/1/102underwater dockingguide positioningYOLO-Dunderwater target detectionvisual positioning |
spellingShingle | Tian Ni Can Sima Wenzhong Zhang Junlin Wang Jia Guo Lindan Zhang Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D Journal of Marine Science and Engineering underwater docking guide positioning YOLO-D underwater target detection visual positioning |
title | Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D |
title_full | Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D |
title_fullStr | Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D |
title_full_unstemmed | Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D |
title_short | Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D |
title_sort | vision based underwater docking guidance and positioning enhancing detection with yolo d |
topic | underwater docking guide positioning YOLO-D underwater target detection visual positioning |
url | https://www.mdpi.com/2077-1312/13/1/102 |
work_keys_str_mv | AT tianni visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod AT cansima visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod AT wenzhongzhang visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod AT junlinwang visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod AT jiaguo visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod AT lindanzhang visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod |