Unmanned aerial vehicle pitch control using deep reinforcement learning with discrete actions in wind tunnel test
Unmanned aerial vehicle pitch control using deep reinforcement learning with discrete actions in wind tunnel test
Deep reinforcement learning is a promising method for training a nonlinear attitude controller for fixed-wing unmanned aerial vehicles. Until now, proof-of-concept studies have demonstrated successful attitude control in simulation. However, detailed experimental investigations have not yet been conducted. This study applied deep reinforcement learning for one-degree-of-freedom pitch control in wind tunnel tests with the aim of gaining practical understandings of attitude control application. Three controllers with different discrete action choices, that is, elevator angles, were designed. The controllers with larger action rates exhibited better performance in terms of following angle-of-attack commands. The root mean square errors for tracking angle-of-attack commands decreased from 3.42° to 1.99° as the maximum action rate increased from 10°/s to 50°/s. The comparison between experimental and simulation results showed that the controller with a smaller action rate experienced the friction effect, and the controllers with larger action rates experienced fluctuating behaviors in elevator maneuvers owing to delay. The investigation of the effect of friction and delay on pitch control highlighted the importance of conducting experiments to understand actual control performances, specifically when the controllers were trained with a low-fidelity model.
Wada, Daichi
20a1527a-c198-4911-bce2-e1d8f13b2dd6
Araujo-Estrada, Sergio A.
87793c63-f2bd-4169-b93d-ec1525909a7a
Windsor, Shane
07554535-35c2-414d-9ad8-e549250da85a
14 January 2021
Wada, Daichi
20a1527a-c198-4911-bce2-e1d8f13b2dd6
Araujo-Estrada, Sergio A.
87793c63-f2bd-4169-b93d-ec1525909a7a
Windsor, Shane
07554535-35c2-414d-9ad8-e549250da85a
Wada, Daichi, Araujo-Estrada, Sergio A. and Windsor, Shane
(2021)
Unmanned aerial vehicle pitch control using deep reinforcement learning with discrete actions in wind tunnel test.
Aerospace, 8 (1).
(doi:10.3390/aerospace8010018).
Abstract
Deep reinforcement learning is a promising method for training a nonlinear attitude controller for fixed-wing unmanned aerial vehicles. Until now, proof-of-concept studies have demonstrated successful attitude control in simulation. However, detailed experimental investigations have not yet been conducted. This study applied deep reinforcement learning for one-degree-of-freedom pitch control in wind tunnel tests with the aim of gaining practical understandings of attitude control application. Three controllers with different discrete action choices, that is, elevator angles, were designed. The controllers with larger action rates exhibited better performance in terms of following angle-of-attack commands. The root mean square errors for tracking angle-of-attack commands decreased from 3.42° to 1.99° as the maximum action rate increased from 10°/s to 50°/s. The comparison between experimental and simulation results showed that the controller with a smaller action rate experienced the friction effect, and the controllers with larger action rates experienced fluctuating behaviors in elevator maneuvers owing to delay. The investigation of the effect of friction and delay on pitch control highlighted the importance of conducting experiments to understand actual control performances, specifically when the controllers were trained with a low-fidelity model.
Text
aerospace-08-00018-v2
- Version of Record
More information
Accepted/In Press date: 8 January 2021
Published date: 14 January 2021
Identifiers
Local EPrints ID: 469047
URI: http://eprints.soton.ac.uk/id/eprint/469047
ISSN: 2226-4310
PURE UUID: 2a94eced-ea71-4434-ba96-0f7037c0c2f3
Catalogue record
Date deposited: 05 Sep 2022 17:02
Last modified: 17 Mar 2024 04:12
Export record
Altmetrics
Contributors
Author:
Daichi Wada
Author:
Sergio A. Araujo-Estrada
Author:
Shane Windsor
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics