Iterative learning control: from model-based to reinforcement learning
Iterative learning control: from model-based to reinforcement learning
High-performance control systems that repeat the same tasks have a wide range of applications. Iterative learning control (ILC) is a control method that enables such systems to achieve high-performance tracking by updating the input based on historical data. Based on whether a system model is required or not, ILC algorithms can be divided into model-based and model-free algorithms. Model-based ILC uses system dynamics (though not necessarily accurate) to update the inputs, whereas model-free ILC only uses input-output data to update. However, model-free ILC techniques typically converge slower than model-based ILC techniques, despite retaining high performance. It is noteworthy that the idea of adapting actions based on past information is also at the core of reinforcement learning (RL). More excitingly, RL provides a number of model-free methods for determining an optimal action. Despite being in different subject areas, RL and ILC have many similarities. This motivates the author to present the first in-depth study of the relationship between ILC and RL from the viewpoint of high-performance tracking problems and propose several novel model-free ILC designs. This thesis starts with a quantitative comparison of ILC and RL techniques in both model-based and model-free scenarios from a control perspective. ILC is shown to be more data-efficient when model information is unavailable, which suggests that research into the structure of the problem may lead to improved performance and opens the door for the development of RL-based ILC (RILC) algorithms. Policy gradient methods and Q-learning from reinforcement learning are used to develop new model-free ILC algorithms. The proposed algorithms achieve high-performance tracking without any model information. Their convergence performance is comparable to their model-based counterparts under certain conditions. Moreover, these algorithms are shown to have the potential to handle nonlinear dynamics. Numerical simulations are presented to demonstrate the effectiveness of the proposed design.
University of Southampton
Zhang, Yueqing
8c1c1b35-d4bd-43d2-8871-b351bcbd2e7c
September 2023
Zhang, Yueqing
8c1c1b35-d4bd-43d2-8871-b351bcbd2e7c
Chu, Bing
555a86a5-0198-4242-8525-3492349d4f0f
Zhang, Yueqing
(2023)
Iterative learning control: from model-based to reinforcement learning.
University of Southampton, Doctoral Thesis, 136pp.
Record type:
Thesis
(Doctoral)
Abstract
High-performance control systems that repeat the same tasks have a wide range of applications. Iterative learning control (ILC) is a control method that enables such systems to achieve high-performance tracking by updating the input based on historical data. Based on whether a system model is required or not, ILC algorithms can be divided into model-based and model-free algorithms. Model-based ILC uses system dynamics (though not necessarily accurate) to update the inputs, whereas model-free ILC only uses input-output data to update. However, model-free ILC techniques typically converge slower than model-based ILC techniques, despite retaining high performance. It is noteworthy that the idea of adapting actions based on past information is also at the core of reinforcement learning (RL). More excitingly, RL provides a number of model-free methods for determining an optimal action. Despite being in different subject areas, RL and ILC have many similarities. This motivates the author to present the first in-depth study of the relationship between ILC and RL from the viewpoint of high-performance tracking problems and propose several novel model-free ILC designs. This thesis starts with a quantitative comparison of ILC and RL techniques in both model-based and model-free scenarios from a control perspective. ILC is shown to be more data-efficient when model information is unavailable, which suggests that research into the structure of the problem may lead to improved performance and opens the door for the development of RL-based ILC (RILC) algorithms. Policy gradient methods and Q-learning from reinforcement learning are used to develop new model-free ILC algorithms. The proposed algorithms achieve high-performance tracking without any model information. Their convergence performance is comparable to their model-based counterparts under certain conditions. Moreover, these algorithms are shown to have the potential to handle nonlinear dynamics. Numerical simulations are presented to demonstrate the effectiveness of the proposed design.
Text
Yueqing_Zhang_Thesis
- Version of Record
Text
Final-thesis-submission-Examination-Miss-Yueqing-Zhang
Restricted to Repository staff only
More information
Published date: September 2023
Identifiers
Local EPrints ID: 481975
URI: http://eprints.soton.ac.uk/id/eprint/481975
PURE UUID: 107fb734-abe1-4490-8b5f-506c4403c1cf
Catalogue record
Date deposited: 14 Sep 2023 16:45
Last modified: 09 Sep 2024 04:01
Export record
Contributors
Author:
Yueqing Zhang
Thesis advisor:
Bing Chu
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics