Stable and safe reinforcement learning via a barrier-Lyapunov actor-critic approach
Stable and safe reinforcement learning via a barrier-Lyapunov actor-critic approach
Reinforcement Learning (RL) has demonstrated impressive performance in various areas such as video games and robotics. However, ensuring safety and stability, which are two critical properties from a control perspective, remains a significant challenge when using RL to control real-world systems. In this paper, we first provide definitions of safety and stability for the RL system, and then combine the Control Barrier Function (CBF) and Control Lyapunov Function (CLF) methods with the actor-critic method in RL to propose a Barrier-Lyapunov Actor-Critic (BLAC) framework which helps maintain the aforementioned safety and stability for the system. In this framework, CBF constraints for safety and CLF constraint for stability are constructed based on the data sampled from the replay buffer, and the augmented Lagrangian method is used to update the parameters of the RL-based controller. Furthermore, an additional backup controller is introduced in case the RL-based controller cannot provide valid control signals w
1320-1325
Zhao, Liqun
3798aa21-6820-4f9f-b5c0-15ebb771f352
Gatsis, Konstantinos
f808d11b-38f1-4a44-ba56-3364d63558d7
Papachristodoulou, Antonis
e3109556-2fc6-4de8-9324-2601777beab6
19 January 2024
Zhao, Liqun
3798aa21-6820-4f9f-b5c0-15ebb771f352
Gatsis, Konstantinos
f808d11b-38f1-4a44-ba56-3364d63558d7
Papachristodoulou, Antonis
e3109556-2fc6-4de8-9324-2601777beab6
Zhao, Liqun, Gatsis, Konstantinos and Papachristodoulou, Antonis
(2024)
Stable and safe reinforcement learning via a barrier-Lyapunov actor-critic approach.
In Proceedings of the IEEE Conference on Decision and Control.
.
(doi:10.1109/CDC49753.2023.10383742).
Record type:
Conference or Workshop Item
(Paper)
Abstract
Reinforcement Learning (RL) has demonstrated impressive performance in various areas such as video games and robotics. However, ensuring safety and stability, which are two critical properties from a control perspective, remains a significant challenge when using RL to control real-world systems. In this paper, we first provide definitions of safety and stability for the RL system, and then combine the Control Barrier Function (CBF) and Control Lyapunov Function (CLF) methods with the actor-critic method in RL to propose a Barrier-Lyapunov Actor-Critic (BLAC) framework which helps maintain the aforementioned safety and stability for the system. In this framework, CBF constraints for safety and CLF constraint for stability are constructed based on the data sampled from the replay buffer, and the augmented Lagrangian method is used to update the parameters of the RL-based controller. Furthermore, an additional backup controller is introduced in case the RL-based controller cannot provide valid control signals w
This record has no associated files available for download.
More information
Published date: 19 January 2024
Identifiers
Local EPrints ID: 494564
URI: http://eprints.soton.ac.uk/id/eprint/494564
PURE UUID: 7d1556ee-1bf7-4e61-881e-ddb90f4dc45f
Catalogue record
Date deposited: 10 Oct 2024 16:46
Last modified: 11 Oct 2024 02:08
Export record
Altmetrics
Contributors
Author:
Liqun Zhao
Author:
Konstantinos Gatsis
Author:
Antonis Papachristodoulou
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics