Slowness-based neural visuomotor control with an intrinsically motivated continuous actor-critic
Slowness-based neural visuomotor control with an intrinsically motivated continuous actor-critic
In this paper, we present a new visually guided exploration approach for
autonomous learning of visuomotor skills. Our approach uses hierarchical Slow
Feature Analysis for unsupervised learning of efficient state representation and an
Intrinsically motivated Continuous Actor-Critic learner for neuro-optimal control.
The system learns online an ensemble of local forward models and generates an
intrinsic reward based on the learning progress of each learned forward model.
Combined with the external reward, the intrinsic reward guides the system’s
exploration strategy. We evaluate the approach for the task of learning to reach an
object using raw pixel data in a realistic robot simulator. The results show that the
control policies learned with our approach are significantly better both in terms of
length and average reward than those learned with any of the baseline algorithms.
509-514
Hafez, Muhammad Burhan
e8c991ab-d800-46f2-abeb-cb169a1ed47e
Kerzel, Matthias
a7ec71f0-3fa1-4acb-a46b-9198ba76ff14
Weber, Cornelius
4e097e6c-840c-460a-8572-e8759f137e43
Wermter, Stefan
80682cc6-4251-420a-af8a-f4d616fb0fcc
25 April 2018
Hafez, Muhammad Burhan
e8c991ab-d800-46f2-abeb-cb169a1ed47e
Kerzel, Matthias
a7ec71f0-3fa1-4acb-a46b-9198ba76ff14
Weber, Cornelius
4e097e6c-840c-460a-8572-e8759f137e43
Wermter, Stefan
80682cc6-4251-420a-af8a-f4d616fb0fcc
Hafez, Muhammad Burhan, Kerzel, Matthias, Weber, Cornelius and Wermter, Stefan
(2018)
Slowness-based neural visuomotor control with an intrinsically motivated continuous actor-critic.
European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, , Bruges, Belgium.
25 - 27 Apr 2018.
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
In this paper, we present a new visually guided exploration approach for
autonomous learning of visuomotor skills. Our approach uses hierarchical Slow
Feature Analysis for unsupervised learning of efficient state representation and an
Intrinsically motivated Continuous Actor-Critic learner for neuro-optimal control.
The system learns online an ensemble of local forward models and generates an
intrinsic reward based on the learning progress of each learned forward model.
Combined with the external reward, the intrinsic reward guides the system’s
exploration strategy. We evaluate the approach for the task of learning to reach an
object using raw pixel data in a realistic robot simulator. The results show that the
control policies learned with our approach are significantly better both in terms of
length and average reward than those learned with any of the baseline algorithms.
Text
h-sfa-based ICAC
- Accepted Manuscript
More information
Published date: 25 April 2018
Venue - Dates:
European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, , Bruges, Belgium, 2018-04-25 - 2018-04-27
Identifiers
Local EPrints ID: 495809
URI: http://eprints.soton.ac.uk/id/eprint/495809
PURE UUID: 9fb60b96-2f51-473a-8d2e-0a7302f750da
Catalogue record
Date deposited: 22 Nov 2024 18:07
Last modified: 23 Nov 2024 03:11
Export record
Contributors
Author:
Muhammad Burhan Hafez
Author:
Matthias Kerzel
Author:
Cornelius Weber
Author:
Stefan Wermter
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics