Parameter optimal iterative learning control design from model-based, data-driven to reinforcement learning
Parameter optimal iterative learning control design from model-based, data-driven to reinforcement learning
Iterative learning control (ILC) is a high-performance control design method for systems operating in a repetitive fashion by learning from past experience. Our recent work shows that reinforcement learning (RL) shares many features with ILC and thus opens the door to new ILC algorithm designs. This paper continues the research by considering a parameter optimal iterative learning control (POILC) algorithm. It has a very simple structure and appealing convergence properties, but requires a model of the system. We first develop a data-driven POILC algorithm without using model information by performing an extra experiment on the plant. We then use a policy gradient RL algorithm to design a new model-free POILC algorithm. Both algorithms achieve the high-performance control target without using model information, but the convergence properties do differ. In particular, by increasing the number of function approximators in the latter, the RL-based model-free ILC can approach the performance of the model-based POILC. A numerical study is presented to compare the performance of different approaches and demonstrate the effectiveness of the proposed designs.
data-based control, Iterative learning control, reinforcement learning control
494-499
Zhang, Yueqing
ab6a3071-e2b6-431e-8b8d-b14e00ade9f6
Chu, Bing
555a86a5-0198-4242-8525-3492349d4f0f
Shu, Zhan
ea5dc18c-d375-4db0-bbcc-dd0229f3a1cb
4 August 2022
Zhang, Yueqing
ab6a3071-e2b6-431e-8b8d-b14e00ade9f6
Chu, Bing
555a86a5-0198-4242-8525-3492349d4f0f
Shu, Zhan
ea5dc18c-d375-4db0-bbcc-dd0229f3a1cb
Zhang, Yueqing, Chu, Bing and Shu, Zhan
(2022)
Parameter optimal iterative learning control design from model-based, data-driven to reinforcement learning.
IFAC-PapersOnLine, 55 (12), .
(doi:10.1016/j.ifacol.2022.07.360).
Abstract
Iterative learning control (ILC) is a high-performance control design method for systems operating in a repetitive fashion by learning from past experience. Our recent work shows that reinforcement learning (RL) shares many features with ILC and thus opens the door to new ILC algorithm designs. This paper continues the research by considering a parameter optimal iterative learning control (POILC) algorithm. It has a very simple structure and appealing convergence properties, but requires a model of the system. We first develop a data-driven POILC algorithm without using model information by performing an extra experiment on the plant. We then use a policy gradient RL algorithm to design a new model-free POILC algorithm. Both algorithms achieve the high-performance control target without using model information, but the convergence properties do differ. In particular, by increasing the number of function approximators in the latter, the RL-based model-free ILC can approach the performance of the model-based POILC. A numerical study is presented to compare the performance of different approaches and demonstrate the effectiveness of the proposed designs.
Text
1-s2.0-S2405896322007613-main
- Version of Record
More information
e-pub ahead of print date: 4 August 2022
Published date: 4 August 2022
Additional Information:
Funding Information:
This work was partially supported by the ZZU-Southampton Collaborative Research Project 16306/01 and the China Scholarship Council (CSC)
Venue - Dates:
14th IFAC Workshop on Adaptive and Learning Control Systems, ALCOS 2022, , Casablanca, Morocco, 2022-06-29 - 2022-07-01
Keywords:
data-based control, Iterative learning control, reinforcement learning control
Identifiers
Local EPrints ID: 471605
URI: http://eprints.soton.ac.uk/id/eprint/471605
ISSN: 2405-8963
PURE UUID: 22cb7a7a-8e7c-4e16-9891-be49367b3dd7
Catalogue record
Date deposited: 14 Nov 2022 18:09
Last modified: 30 Sep 2025 02:19
Export record
Altmetrics
Contributors
Author:
Yueqing Zhang
Author:
Bing Chu
Author:
Zhan Shu
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics