The University of Southampton
University of Southampton Institutional Repository

Reinforced continual learning

Reinforced continual learning
Reinforced continual learning

Most artificial intelligence models are limited in their ability to solve new tasks faster, without forgetting previously acquired knowledge. The recently emerging paradigm of continual learning aims to solve this issue, in which the model learns various tasks in a sequential fashion. In this work, a novel approach for continual learning is proposed, which searches for the best neural architecture for each coming task via sophisticatedly designed reinforcement learning strategies. We name it as Reinforced Continual Learning. Our method not only has good performance on preventing catastrophic forgetting but also fits new tasks well. The experiments on sequential classification tasks for variants of MNIST and CIFAR-100 datasets demonstrate that the proposed approach outperforms existing continual learning alternatives for deep networks.

899-908
Xu, Ju
01a01bbe-ba4e-43f8-a421-9d5ef5b25f9a
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
Xu, Ju
01a01bbe-ba4e-43f8-a421-9d5ef5b25f9a
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0

Xu, Ju and Zhu, Zhanxing (2018) Reinforced continual learning. 32nd Conference on Neural Information Processing Systems, Palais des Congrès de Montréal, Montréal, Canada. 02 - 08 Dec 2018. pp. 899-908 .

Record type: Conference or Workshop Item (Paper)

Abstract

Most artificial intelligence models are limited in their ability to solve new tasks faster, without forgetting previously acquired knowledge. The recently emerging paradigm of continual learning aims to solve this issue, in which the model learns various tasks in a sequential fashion. In this work, a novel approach for continual learning is proposed, which searches for the best neural architecture for each coming task via sophisticatedly designed reinforcement learning strategies. We name it as Reinforced Continual Learning. Our method not only has good performance on preventing catastrophic forgetting but also fits new tasks well. The experiments on sequential classification tasks for variants of MNIST and CIFAR-100 datasets demonstrate that the proposed approach outperforms existing continual learning alternatives for deep networks.

This record has no associated files available for download.

More information

Published date: 2018
Additional Information: Funding Information: Supported by National Natural Science Foundation of China (Grant No: 61806009) and Beijing Natural Science Foundation (Grant No: 4184090). Publisher Copyright: © 2018 Curran Associates Inc..All rights reserved.
Venue - Dates: 32nd Conference on Neural Information Processing Systems, Palais des Congrès de Montréal, Montréal, Canada, 2018-12-02 - 2018-12-08

Identifiers

Local EPrints ID: 486128
URI: http://eprints.soton.ac.uk/id/eprint/486128
PURE UUID: 5de306c6-466b-40f8-a434-e5aff6e76039

Catalogue record

Date deposited: 10 Jan 2024 17:38
Last modified: 17 Mar 2024 13:43

Export record

Contributors

Author: Ju Xu
Author: Zhanxing Zhu

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×