A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments
A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments
Unmanned aerial vehicles (UAVs) can be considered in many applications, such as wireless communication, logistics transportation, agriculture and disaster prevention. The flexible maneuverability of UAVs also means that the UAV often operates in complex 3D environments, which requires efficient and reliable path planning system support. However, as a limited resource platform, the UAV systems cannot support highly complex path planning algorithms in lots of scenarios. In this paper, we propose a 3D spatial information compression (3DSIC) based deep reinforcement learning (DRL) algorithm for UAV path planning in cluttered 3D environments. Specifically, the proposed algorithm compresses the 3D spatial information to 2D through 3DSIC, and then combines the compressed 2D environment information with the current UAV layer spatial information to train UAV agents for path planning using neural networks. Additionally, the proposed 3DSIC is a plug and use module that can be combined with various
DRL frameworks such as Deep Q-Network (DQN) and deep deterministic policy gradient (DDPG). Our simulation results show that the training efficiency of 3DSIC-DQN is 4.028 times higher than that directly implementing DQN in a 100×100×50 3D cluttered environment. Furthermore, the training efficiency of 3DSIC-DDPG is 3.9 times higher than the traditional DDPG in the same environment. Moreover, 3DSIC combined with fast recurrent stochastic value gradient (FRSVG), which can be considered as the most state-of-the-art DRL algorithm for UAV path planning, exhibits 2.35 times faster training speed compared with the original FRSVG algorithm.
unmanned aerial vehicles, 3D spatial information compression, deep reinforcement learning, training efficiency, 3D path planning
647-661
Wang, Zhipeng
feb79a9c-caba-4f0c-a561-dff6447aae64
Ng, Soon Xin
e19a63b0-0f12-4591-ab5f-554820d5f78c
El-Hajjar, Mohammed
3a829028-a427-4123-b885-2bab81a44b6f
10 February 2025
Wang, Zhipeng
feb79a9c-caba-4f0c-a561-dff6447aae64
Ng, Soon Xin
e19a63b0-0f12-4591-ab5f-554820d5f78c
El-Hajjar, Mohammed
3a829028-a427-4123-b885-2bab81a44b6f
Wang, Zhipeng, Ng, Soon Xin and El-Hajjar, Mohammed
(2025)
A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments.
IEEE Open Journal of Vehicular Technology, 6, .
(doi:10.1109/OJVT.2025.3540174).
Abstract
Unmanned aerial vehicles (UAVs) can be considered in many applications, such as wireless communication, logistics transportation, agriculture and disaster prevention. The flexible maneuverability of UAVs also means that the UAV often operates in complex 3D environments, which requires efficient and reliable path planning system support. However, as a limited resource platform, the UAV systems cannot support highly complex path planning algorithms in lots of scenarios. In this paper, we propose a 3D spatial information compression (3DSIC) based deep reinforcement learning (DRL) algorithm for UAV path planning in cluttered 3D environments. Specifically, the proposed algorithm compresses the 3D spatial information to 2D through 3DSIC, and then combines the compressed 2D environment information with the current UAV layer spatial information to train UAV agents for path planning using neural networks. Additionally, the proposed 3DSIC is a plug and use module that can be combined with various
DRL frameworks such as Deep Q-Network (DQN) and deep deterministic policy gradient (DDPG). Our simulation results show that the training efficiency of 3DSIC-DQN is 4.028 times higher than that directly implementing DQN in a 100×100×50 3D cluttered environment. Furthermore, the training efficiency of 3DSIC-DDPG is 3.9 times higher than the traditional DDPG in the same environment. Moreover, 3DSIC combined with fast recurrent stochastic value gradient (FRSVG), which can be considered as the most state-of-the-art DRL algorithm for UAV path planning, exhibits 2.35 times faster training speed compared with the original FRSVG algorithm.
Text
A_3D_Spatial_Information_Compression_Based_Deep_Reinforcement_Learning_Technique_for_UAV_Path_Planning_in_Cluttered_Environments
- Version of Record
More information
Accepted/In Press date: 6 February 2025
e-pub ahead of print date: 10 February 2025
Published date: 10 February 2025
Keywords:
unmanned aerial vehicles, 3D spatial information compression, deep reinforcement learning, training efficiency, 3D path planning
Identifiers
Local EPrints ID: 498808
URI: http://eprints.soton.ac.uk/id/eprint/498808
ISSN: 2644-1330
PURE UUID: 62dc80a5-d023-4874-8cb1-9506cb2f0a55
Catalogue record
Date deposited: 28 Feb 2025 18:00
Last modified: 22 Aug 2025 02:32
Export record
Altmetrics
Contributors
Author:
Zhipeng Wang
Author:
Soon Xin Ng
Author:
Mohammed El-Hajjar
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics