OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments

UAVs typically rely on satellite navigation for positioning, yet this method proves ineffective in instances where the signal is inadequate or communication is disrupted. Visually based positioning technology has emerged as a reliable alternative. In this paper, we propose a novel end-to-end network...

Full description

Saved in:
Bibliographic Details
Main Authors: Nanxing Chen, Jiqi Fan, Jiayu Yuan, Enhui Zheng
Format: Article
Language:English
Published: MDPI AG 2025-01-01
Series:Drones
Subjects:
Online Access:https://www.mdpi.com/2504-446X/9/1/33
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832588648957083648
author Nanxing Chen
Jiqi Fan
Jiayu Yuan
Enhui Zheng
author_facet Nanxing Chen
Jiqi Fan
Jiayu Yuan
Enhui Zheng
author_sort Nanxing Chen
collection DOAJ
description UAVs typically rely on satellite navigation for positioning, yet this method proves ineffective in instances where the signal is inadequate or communication is disrupted. Visually based positioning technology has emerged as a reliable alternative. In this paper, we propose a novel end-to-end network, OBTPN. In the initial phase of the model, we optimized the distribution of attention within the primary network, aiming to achieve a balance between self-attention and cross-attention. Subsequently, we devised a feature fusion head, which enhanced the model’s capacity to process multi-scale information. OBTPN was successfully deployed on an NVIDIA Jetson TX2 onboard computer. This paper also proposes a high-altitude complex environment dataset, Crossview9, which addresses a research gap in the field of high-altitude visual navigation. The performance of the model on this dataset is also evaluated. Additionally, the dataset was processed to simulate a low-quality image environment to assess the model’s resilience in challenging weather conditions. The experimental results demonstrate that OBTPN_256 attains an accuracy of 84.55% on the RDS metric, thereby reaching the state-of-the-art (SOTA) level of the UL14 dataset. On the Crossview9 dataset, OBTPN_256 achieves an RDS score of 79.76%, also reaching the SOTA level. Most notably, the model’s high accuracy in low-quality image environments further substantiates its robustness in complex environments.
format Article
id doaj-art-c3fa90b351dd4affa5625fe6151528d6
institution Kabale University
issn 2504-446X
language English
publishDate 2025-01-01
publisher MDPI AG
record_format Article
series Drones
spelling doaj-art-c3fa90b351dd4affa5625fe6151528d62025-01-24T13:29:43ZengMDPI AGDrones2504-446X2025-01-01913310.3390/drones9010033OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude EnvironmentsNanxing Chen0Jiqi Fan1Jiayu Yuan2Enhui Zheng3School of Mechanical and Electrical Engineering, China Jiliang University, Hangzhou 310018, ChinaSchool of Mechanical and Electrical Engineering, China Jiliang University, Hangzhou 310018, ChinaSchool of Mechanical and Electrical Engineering, China Jiliang University, Hangzhou 310018, ChinaSchool of Mechanical and Electrical Engineering, China Jiliang University, Hangzhou 310018, ChinaUAVs typically rely on satellite navigation for positioning, yet this method proves ineffective in instances where the signal is inadequate or communication is disrupted. Visually based positioning technology has emerged as a reliable alternative. In this paper, we propose a novel end-to-end network, OBTPN. In the initial phase of the model, we optimized the distribution of attention within the primary network, aiming to achieve a balance between self-attention and cross-attention. Subsequently, we devised a feature fusion head, which enhanced the model’s capacity to process multi-scale information. OBTPN was successfully deployed on an NVIDIA Jetson TX2 onboard computer. This paper also proposes a high-altitude complex environment dataset, Crossview9, which addresses a research gap in the field of high-altitude visual navigation. The performance of the model on this dataset is also evaluated. Additionally, the dataset was processed to simulate a low-quality image environment to assess the model’s resilience in challenging weather conditions. The experimental results demonstrate that OBTPN_256 attains an accuracy of 84.55% on the RDS metric, thereby reaching the state-of-the-art (SOTA) level of the UL14 dataset. On the Crossview9 dataset, OBTPN_256 achieves an RDS score of 79.76%, also reaching the SOTA level. Most notably, the model’s high accuracy in low-quality image environments further substantiates its robustness in complex environments.https://www.mdpi.com/2504-446X/9/1/33UAVgeo-localizationtransformerhigh-altitude datasetonboard deployment
spellingShingle Nanxing Chen
Jiqi Fan
Jiayu Yuan
Enhui Zheng
OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments
Drones
UAV
geo-localization
transformer
high-altitude dataset
onboard deployment
title OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments
title_full OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments
title_fullStr OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments
title_full_unstemmed OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments
title_short OBTPN: A Vision-Based Network for UAV Geo-Localization in Multi-Altitude Environments
title_sort obtpn a vision based network for uav geo localization in multi altitude environments
topic UAV
geo-localization
transformer
high-altitude dataset
onboard deployment
url https://www.mdpi.com/2504-446X/9/1/33
work_keys_str_mv AT nanxingchen obtpnavisionbasednetworkforuavgeolocalizationinmultialtitudeenvironments
AT jiqifan obtpnavisionbasednetworkforuavgeolocalizationinmultialtitudeenvironments
AT jiayuyuan obtpnavisionbasednetworkforuavgeolocalizationinmultialtitudeenvironments
AT enhuizheng obtpnavisionbasednetworkforuavgeolocalizationinmultialtitudeenvironments