UAV Collision Avoidance in Unknown Scenarios with Causal Representation Disentanglement

Deep reinforcement learning (DRL) has significantly advanced online path planning for unmanned aerial vehicles (UAVs). Nonetheless, DRL-based methods often encounter reduced performance when dealing with unfamiliar scenarios. This decline is mainly caused by the presence of non-causal and domain-spe...

Full description

Saved in:
Bibliographic Details
Main Authors: Zhun Fan, Zihao Xia, Che Lin, Gaofei Han, Wenji Li, Dongliang Wang, Yindong Chen, Zhifeng Hao, Ruichu Cai, Jiafan Zhuang
Format: Article
Language:English
Published: MDPI AG 2024-12-01
Series:Drones
Subjects:
Online Access:https://www.mdpi.com/2504-446X/9/1/10
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Deep reinforcement learning (DRL) has significantly advanced online path planning for unmanned aerial vehicles (UAVs). Nonetheless, DRL-based methods often encounter reduced performance when dealing with unfamiliar scenarios. This decline is mainly caused by the presence of non-causal and domain-specific elements within visual representations, which negatively impact the learning of policies. Present techniques generally depend on predefined augmentation or regularization methods intended to direct the model toward identifying causal and domain-invariant components, thereby enhancing the model’s ability to generalize. However, these manually crafted approaches are intrinsically constrained in their coverage and do not consider the entire spectrum of possible scenarios, resulting in less effective performance in new environments. Unlike prior studies, this work prioritizes representation learning and presents a novel method for causal representation disentanglement. The approach successfully distinguishes between causal and non-causal elements in visual data. By concentrating on causal aspects during the policy learning phase, the impact of non-causal factors is minimized, thereby improving the generalizability of DRL models. Experimental results demonstrate that our technique achieves reliable navigation and effective collision avoidance in unseen scenarios, surpassing state-of-the-art deep reinforcement learning algorithms.
ISSN:2504-446X