The University of Southampton
University of Southampton Institutional Repository

Federated reinforcement learning at the edge: exploring the learning-communication tradeoff

Federated reinforcement learning at the edge: exploring the learning-communication tradeoff
Federated reinforcement learning at the edge: exploring the learning-communication tradeoff

Modern cyber-physical architectures use data col-lected from systems at different physical locations to learn appropriate behaviors and adapt to uncertain environments. However, an important challenge arises as communication exchanges at the edge of networked systems are costly due to limited resources. This paper considers a setup where multiple agents need to communicate efficiently in order to jointly solve a reinforcement learning problem over time-series data collected in a distributed manner. This is posed as learning an approximate value function over a communication network. An algorithm for achieving communication efficiency is proposed, supported with theoretical guarantees, practical implementations, and numerical evaluations. The approach is based on the idea of communicating only when sufficiently informative data is collected.

1890-1895
IEEE
Gatsis, Konstantinos
f808d11b-38f1-4a44-ba56-3364d63558d7
Gatsis, Konstantinos
f808d11b-38f1-4a44-ba56-3364d63558d7

Gatsis, Konstantinos (2022) Federated reinforcement learning at the edge: exploring the learning-communication tradeoff. In 2022 European Control Conference, ECC 2022. IEEE. pp. 1890-1895 . (doi:10.23919/ECC55457.2022.9837987).

Record type: Conference or Workshop Item (Paper)

Abstract

Modern cyber-physical architectures use data col-lected from systems at different physical locations to learn appropriate behaviors and adapt to uncertain environments. However, an important challenge arises as communication exchanges at the edge of networked systems are costly due to limited resources. This paper considers a setup where multiple agents need to communicate efficiently in order to jointly solve a reinforcement learning problem over time-series data collected in a distributed manner. This is posed as learning an approximate value function over a communication network. An algorithm for achieving communication efficiency is proposed, supported with theoretical guarantees, practical implementations, and numerical evaluations. The approach is based on the idea of communicating only when sufficiently informative data is collected.

This record has no associated files available for download.

More information

Published date: 12 July 2022
Additional Information: Publisher Copyright: © 2022 EUCA.
Venue - Dates: 2022 European Control Conference, ECC 2022, , London, United Kingdom, 2022-07-12 - 2022-07-15

Identifiers

Local EPrints ID: 494234
URI: http://eprints.soton.ac.uk/id/eprint/494234
PURE UUID: d2e81a50-c16e-40f1-b9c4-c08eaa7644fe
ORCID for Konstantinos Gatsis: ORCID iD orcid.org/0000-0002-0734-5445

Catalogue record

Date deposited: 01 Oct 2024 16:50
Last modified: 02 Oct 2024 02:12

Export record

Altmetrics

Contributors

Author: Konstantinos Gatsis ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×