Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D

This study proposed a vision-based underwater vertical docking guidance and positioning method to address docking control challenges for human-operated vehicles (HOVs) and unmanned underwater vehicles (UUVs) under complex underwater visual conditions. A cascaded detection and positioning strategy in...

Full description

Saved in:
Bibliographic Details
Main Authors: Tian Ni, Can Sima, Wenzhong Zhang, Junlin Wang, Jia Guo, Lindan Zhang
Format: Article
Language:English
Published: MDPI AG 2025-01-01
Series:Journal of Marine Science and Engineering
Subjects:
Online Access:https://www.mdpi.com/2077-1312/13/1/102
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832588232299118592
author Tian Ni
Can Sima
Wenzhong Zhang
Junlin Wang
Jia Guo
Lindan Zhang
author_facet Tian Ni
Can Sima
Wenzhong Zhang
Junlin Wang
Jia Guo
Lindan Zhang
author_sort Tian Ni
collection DOAJ
description This study proposed a vision-based underwater vertical docking guidance and positioning method to address docking control challenges for human-operated vehicles (HOVs) and unmanned underwater vehicles (UUVs) under complex underwater visual conditions. A cascaded detection and positioning strategy incorporating fused active and passive markers enabled real-time detection of the relative position and pose between the UUV and docking station (DS). A novel deep learning-based network model, YOLO-D, was developed to detect docking markers in real time. YOLO-D employed the Adaptive Kernel Convolution Module (AKConv) to dynamically adjust the sample shapes and sizes and optimize the target feature detection across various scales and regions. It integrated the Context Aggregation Network (CONTAINER) to enhance small-target detection and overall image accuracy, while the bidirectional feature pyramid network (BiFPN) facilitated effective cross-scale feature fusion, improving detection precision for multi-scale and fuzzy targets. In addition, an underwater docking positioning algorithm leveraging multiple markers was implemented. Tests on an underwater docking markers dataset demonstrated that YOLO-D achieved a detection accuracy of mAP@0.5 to 94.5%, surpassing the baseline YOLOv11n with improvements of 1.5% in precision, 5% in recall, and 4.2% in mAP@0.5. Pool experiments verified the feasibility of the method, achieving a 90% success rate for single-attempt docking and recovery. The proposed approach offered an accurate and efficient solution for underwater docking guidance and target detection, which is of great significance for improving the safety of docking.
format Article
id doaj-art-c6f48eb94638434aac54800ce18b4b6f
institution Kabale University
issn 2077-1312
language English
publishDate 2025-01-01
publisher MDPI AG
record_format Article
series Journal of Marine Science and Engineering
spelling doaj-art-c6f48eb94638434aac54800ce18b4b6f2025-01-24T13:36:51ZengMDPI AGJournal of Marine Science and Engineering2077-13122025-01-0113110210.3390/jmse13010102Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-DTian Ni0Can Sima1Wenzhong Zhang2Junlin Wang3Jia Guo4Lindan Zhang5China Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaChina Ship Scientific Research Centre, Wuxi 214082, ChinaThis study proposed a vision-based underwater vertical docking guidance and positioning method to address docking control challenges for human-operated vehicles (HOVs) and unmanned underwater vehicles (UUVs) under complex underwater visual conditions. A cascaded detection and positioning strategy incorporating fused active and passive markers enabled real-time detection of the relative position and pose between the UUV and docking station (DS). A novel deep learning-based network model, YOLO-D, was developed to detect docking markers in real time. YOLO-D employed the Adaptive Kernel Convolution Module (AKConv) to dynamically adjust the sample shapes and sizes and optimize the target feature detection across various scales and regions. It integrated the Context Aggregation Network (CONTAINER) to enhance small-target detection and overall image accuracy, while the bidirectional feature pyramid network (BiFPN) facilitated effective cross-scale feature fusion, improving detection precision for multi-scale and fuzzy targets. In addition, an underwater docking positioning algorithm leveraging multiple markers was implemented. Tests on an underwater docking markers dataset demonstrated that YOLO-D achieved a detection accuracy of mAP@0.5 to 94.5%, surpassing the baseline YOLOv11n with improvements of 1.5% in precision, 5% in recall, and 4.2% in mAP@0.5. Pool experiments verified the feasibility of the method, achieving a 90% success rate for single-attempt docking and recovery. The proposed approach offered an accurate and efficient solution for underwater docking guidance and target detection, which is of great significance for improving the safety of docking.https://www.mdpi.com/2077-1312/13/1/102underwater dockingguide positioningYOLO-Dunderwater target detectionvisual positioning
spellingShingle Tian Ni
Can Sima
Wenzhong Zhang
Junlin Wang
Jia Guo
Lindan Zhang
Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
Journal of Marine Science and Engineering
underwater docking
guide positioning
YOLO-D
underwater target detection
visual positioning
title Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
title_full Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
title_fullStr Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
title_full_unstemmed Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
title_short Vision-Based Underwater Docking Guidance and Positioning: Enhancing Detection with YOLO-D
title_sort vision based underwater docking guidance and positioning enhancing detection with yolo d
topic underwater docking
guide positioning
YOLO-D
underwater target detection
visual positioning
url https://www.mdpi.com/2077-1312/13/1/102
work_keys_str_mv AT tianni visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod
AT cansima visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod
AT wenzhongzhang visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod
AT junlinwang visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod
AT jiaguo visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod
AT lindanzhang visionbasedunderwaterdockingguidanceandpositioningenhancingdetectionwithyolod