Reinforced continual learning
Reinforced continual learning
Most artificial intelligence models are limited in their ability to solve new tasks faster, without forgetting previously acquired knowledge. The recently emerging paradigm of continual learning aims to solve this issue, in which the model learns various tasks in a sequential fashion. In this work, a novel approach for continual learning is proposed, which searches for the best neural architecture for each coming task via sophisticatedly designed reinforcement learning strategies. We name it as Reinforced Continual Learning. Our method not only has good performance on preventing catastrophic forgetting but also fits new tasks well. The experiments on sequential classification tasks for variants of MNIST and CIFAR-100 datasets demonstrate that the proposed approach outperforms existing continual learning alternatives for deep networks.
899-908
Xu, Ju
01a01bbe-ba4e-43f8-a421-9d5ef5b25f9a
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
2018
Xu, Ju
01a01bbe-ba4e-43f8-a421-9d5ef5b25f9a
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
Xu, Ju and Zhu, Zhanxing
(2018)
Reinforced continual learning.
32nd Conference on Neural Information Processing Systems, Palais des Congrès de Montréal, Montréal, Canada.
02 - 08 Dec 2018.
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
Most artificial intelligence models are limited in their ability to solve new tasks faster, without forgetting previously acquired knowledge. The recently emerging paradigm of continual learning aims to solve this issue, in which the model learns various tasks in a sequential fashion. In this work, a novel approach for continual learning is proposed, which searches for the best neural architecture for each coming task via sophisticatedly designed reinforcement learning strategies. We name it as Reinforced Continual Learning. Our method not only has good performance on preventing catastrophic forgetting but also fits new tasks well. The experiments on sequential classification tasks for variants of MNIST and CIFAR-100 datasets demonstrate that the proposed approach outperforms existing continual learning alternatives for deep networks.
This record has no associated files available for download.
More information
Published date: 2018
Additional Information:
Funding Information:
Supported by National Natural Science Foundation of China (Grant No: 61806009) and Beijing Natural Science Foundation (Grant No: 4184090).
Publisher Copyright:
© 2018 Curran Associates Inc..All rights reserved.
Venue - Dates:
32nd Conference on Neural Information Processing Systems, Palais des Congrès de Montréal, Montréal, Canada, 2018-12-02 - 2018-12-08
Identifiers
Local EPrints ID: 486128
URI: http://eprints.soton.ac.uk/id/eprint/486128
PURE UUID: 5de306c6-466b-40f8-a434-e5aff6e76039
Catalogue record
Date deposited: 10 Jan 2024 17:38
Last modified: 17 Mar 2024 13:43
Export record
Contributors
Author:
Ju Xu
Author:
Zhanxing Zhu
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics