The University of Southampton
University of Southampton Institutional Repository

A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments

A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments
A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments
Unmanned aerial vehicles (UAVs) can be considered in many applications, such as wireless communication, logistics transportation, agriculture and disaster prevention. The flexible maneuverability of UAVs also means that the UAV often operates in complex 3D environments, which requires efficient and reliable path planning system support. However, as a limited resource platform, the UAV systems cannot support highly complex path planning algorithms in lots of scenarios. In this paper, we propose a 3D spatial information compression (3DSIC) based deep reinforcement learning (DRL) algorithm for UAV path planning in cluttered 3D environments. Specifically, the proposed algorithm compresses the 3D spatial information to 2D through 3DSIC, and then combines the compressed 2D environment information with the current UAV layer spatial information to train UAV agents for path planning using neural networks. Additionally, the proposed 3DSIC is a plug and use module that can be combined with various
DRL frameworks such as Deep Q-Network (DQN) and deep deterministic policy gradient (DDPG). Our simulation results show that the training efficiency of 3DSIC-DQN is 4.028 times higher than that directly implementing DQN in a 100×100×50 3D cluttered environment. Furthermore, the training efficiency of 3DSIC-DDPG is 3.9 times higher than the traditional DDPG in the same environment. Moreover, 3DSIC combined with fast recurrent stochastic value gradient (FRSVG), which can be considered as the most state-of-the-art DRL algorithm for UAV path planning, exhibits 2.35 times faster training speed compared with the original FRSVG algorithm.
unmanned aerial vehicles, 3D spatial information compression, deep reinforcement learning, training efficiency, 3D path planning
2644-1330
647-661
Wang, Zhipeng
feb79a9c-caba-4f0c-a561-dff6447aae64
Ng, Soon Xin
e19a63b0-0f12-4591-ab5f-554820d5f78c
El-Hajjar, Mohammed
3a829028-a427-4123-b885-2bab81a44b6f
Wang, Zhipeng
feb79a9c-caba-4f0c-a561-dff6447aae64
Ng, Soon Xin
e19a63b0-0f12-4591-ab5f-554820d5f78c
El-Hajjar, Mohammed
3a829028-a427-4123-b885-2bab81a44b6f

Wang, Zhipeng, Ng, Soon Xin and El-Hajjar, Mohammed (2025) A 3D spatial information compression based deep reinforcement learning technique for UAV path planning in cluttered environments. IEEE Open Journal of Vehicular Technology, 6, 647-661. (doi:10.1109/OJVT.2025.3540174).

Record type: Article

Abstract

Unmanned aerial vehicles (UAVs) can be considered in many applications, such as wireless communication, logistics transportation, agriculture and disaster prevention. The flexible maneuverability of UAVs also means that the UAV often operates in complex 3D environments, which requires efficient and reliable path planning system support. However, as a limited resource platform, the UAV systems cannot support highly complex path planning algorithms in lots of scenarios. In this paper, we propose a 3D spatial information compression (3DSIC) based deep reinforcement learning (DRL) algorithm for UAV path planning in cluttered 3D environments. Specifically, the proposed algorithm compresses the 3D spatial information to 2D through 3DSIC, and then combines the compressed 2D environment information with the current UAV layer spatial information to train UAV agents for path planning using neural networks. Additionally, the proposed 3DSIC is a plug and use module that can be combined with various
DRL frameworks such as Deep Q-Network (DQN) and deep deterministic policy gradient (DDPG). Our simulation results show that the training efficiency of 3DSIC-DQN is 4.028 times higher than that directly implementing DQN in a 100×100×50 3D cluttered environment. Furthermore, the training efficiency of 3DSIC-DDPG is 3.9 times higher than the traditional DDPG in the same environment. Moreover, 3DSIC combined with fast recurrent stochastic value gradient (FRSVG), which can be considered as the most state-of-the-art DRL algorithm for UAV path planning, exhibits 2.35 times faster training speed compared with the original FRSVG algorithm.

Text
A_3D_Spatial_Information_Compression_Based_Deep_Reinforcement_Learning_Technique_for_UAV_Path_Planning_in_Cluttered_Environments - Version of Record
Available under License Creative Commons Attribution.
Download (4MB)
Text
paper - Proof
Available under License Creative Commons Attribution.
Download (4MB)

More information

Accepted/In Press date: 6 February 2025
e-pub ahead of print date: 10 February 2025
Published date: 10 February 2025
Keywords: unmanned aerial vehicles, 3D spatial information compression, deep reinforcement learning, training efficiency, 3D path planning

Identifiers

Local EPrints ID: 498808
URI: http://eprints.soton.ac.uk/id/eprint/498808
ISSN: 2644-1330
PURE UUID: 62dc80a5-d023-4874-8cb1-9506cb2f0a55
ORCID for Zhipeng Wang: ORCID iD orcid.org/0009-0004-1940-1047
ORCID for Soon Xin Ng: ORCID iD orcid.org/0000-0002-0930-7194
ORCID for Mohammed El-Hajjar: ORCID iD orcid.org/0000-0002-7987-1401

Catalogue record

Date deposited: 28 Feb 2025 18:00
Last modified: 22 Aug 2025 02:32

Export record

Altmetrics

Contributors

Author: Zhipeng Wang ORCID iD
Author: Soon Xin Ng ORCID iD
Author: Mohammed El-Hajjar ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×