Adaptive progressive continual learning
Adaptive progressive continual learning
Continual learning paradigm learns from a continuous stream of tasks in an incremental manner and aims to overcome the notorious issue: the catastrophic forgetting. In this work, we propose a new adaptive progressive network framework including two models for continual learning: Reinforced Continual Learning (RCL) and Bayesian Optimized Continual Learning with Attention mechanism (BOCL) to solve this fundamental issue. The core idea of this framework is to dynamically and adaptively expand the neural network structure upon the arrival of new tasks. RCL and BOCL employ reinforcement learning and Bayesian optimization to achieve it, respectively. An outstanding advantage of our proposed framework is that it will not forget the knowledge that has been learned through adaptively controlling the architecture. We propose effective ways of employing the learned knowledge in the two methods to control the size of the network. RCL employs previous knowledge directly while BOCL selectively utilizes previous knowledge (e.g., feature maps of previous tasks) via attention mechanism. The experiments on variants of MNIST, CIFAR-100 and Sequence of 5-Datasets demonstrate that our methods outperform the state-of-the-art in preventing catastrophic forgetting and fitting new tasks better under the same or less computing resource.
6715-6728
Xu, Ju
01a01bbe-ba4e-43f8-a421-9d5ef5b25f9a
Ma, Jin
67a9490a-b40b-4912-8b07-5b68b4b0fab7
Gao, Xuesong
1363ffd7-6f13-4a22-bd01-a4a0a80ee36d
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
1 October 2021
Xu, Ju
01a01bbe-ba4e-43f8-a421-9d5ef5b25f9a
Ma, Jin
67a9490a-b40b-4912-8b07-5b68b4b0fab7
Gao, Xuesong
1363ffd7-6f13-4a22-bd01-a4a0a80ee36d
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
Xu, Ju, Ma, Jin, Gao, Xuesong and Zhu, Zhanxing
(2021)
Adaptive progressive continual learning.
IEEE Trans. on Pattern Analysis and Machine Intelligence, 44 (10), .
(doi:10.1109/TPAMI.2021.3095064).
Abstract
Continual learning paradigm learns from a continuous stream of tasks in an incremental manner and aims to overcome the notorious issue: the catastrophic forgetting. In this work, we propose a new adaptive progressive network framework including two models for continual learning: Reinforced Continual Learning (RCL) and Bayesian Optimized Continual Learning with Attention mechanism (BOCL) to solve this fundamental issue. The core idea of this framework is to dynamically and adaptively expand the neural network structure upon the arrival of new tasks. RCL and BOCL employ reinforcement learning and Bayesian optimization to achieve it, respectively. An outstanding advantage of our proposed framework is that it will not forget the knowledge that has been learned through adaptively controlling the architecture. We propose effective ways of employing the learned knowledge in the two methods to control the size of the network. RCL employs previous knowledge directly while BOCL selectively utilizes previous knowledge (e.g., feature maps of previous tasks) via attention mechanism. The experiments on variants of MNIST, CIFAR-100 and Sequence of 5-Datasets demonstrate that our methods outperform the state-of-the-art in preventing catastrophic forgetting and fitting new tasks better under the same or less computing resource.
This record has no associated files available for download.
More information
Accepted/In Press date: 18 June 2021
e-pub ahead of print date: 7 July 2021
Published date: 1 October 2021
Identifiers
Local EPrints ID: 485995
URI: http://eprints.soton.ac.uk/id/eprint/485995
PURE UUID: 0524bafc-3839-417d-b586-90e7a9a914a9
Catalogue record
Date deposited: 05 Jan 2024 17:30
Last modified: 17 Oct 2024 16:33
Export record
Altmetrics
Contributors
Author:
Ju Xu
Author:
Jin Ma
Author:
Xuesong Gao
Author:
Zhanxing Zhu
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics