The University of Southampton
University of Southampton Institutional Repository

FPGA acceleration of dynamic neural networks: challenges and advancements

FPGA acceleration of dynamic neural networks: challenges and advancements
FPGA acceleration of dynamic neural networks: challenges and advancements
Modern machine learning methods continue to produce models with a high memory footprint and computational complexity that are increasingly difficult to deploy in resource constrained environments. This is, in part, driven by a focus on costly, power-intensive GPUs, which has a feedback effect on the variety of methods and models chosen for development. We advocate for a transition away from the general purpose processing towards a more targeted, power-efficient, form of hardware, the Field-Programmable Gate Array (FPGA). These devices allow the user to programmatically tailor the model processing architecture, resulting in increased inference performance and lower power demands. Their resources however are limited, which leads to the necessity of simplifying the target deep machine learning models. Dynamic Deep Neural Networks (DNNs) are a class of models that go beyond limits of static model compression, by tuning computational workload to the difficultly of inputs on a per-sample basis. In spite of the model simplification capabilities of Dynamic DNNs and the provable efficiency of FPGAs, little work has been done towards accelerating Dynamic DNNs on FPGAs. In this paper we discuss why this occurs by highlighting the challenges and limitations, both at the software and hardware level. We detail the available efficiency, performance gains, and practical benefits of state-of-the-art Dynamic DNN implementations when FPGAs are adopted as the acceleration device. Finally, we present our conclusions and recommendations for continued research in this space
Dimitriou, Anastasios
02f87799-17dc-4271-96c3-8b30e64e659e
Biggs, Benjamin
8933978b-da66-4f37-bc72-58e41137d940
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
Merrett, Geoff V.
89b3a696-41de-44c3-89aa-b0aa29f54020
Dimitriou, Anastasios
02f87799-17dc-4271-96c3-8b30e64e659e
Biggs, Benjamin
8933978b-da66-4f37-bc72-58e41137d940
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
Merrett, Geoff V.
89b3a696-41de-44c3-89aa-b0aa29f54020

Dimitriou, Anastasios, Biggs, Benjamin, Hare, Jonathon and Merrett, Geoff V. (2024) FPGA acceleration of dynamic neural networks: challenges and advancements. In Proceedings of 2024 IEEE International Conference on Omni-layer Intelligent Systems (COINS). 6 pp . (doi:10.1109/COINS61597.2024.10622127).

Record type: Conference or Workshop Item (Paper)

Abstract

Modern machine learning methods continue to produce models with a high memory footprint and computational complexity that are increasingly difficult to deploy in resource constrained environments. This is, in part, driven by a focus on costly, power-intensive GPUs, which has a feedback effect on the variety of methods and models chosen for development. We advocate for a transition away from the general purpose processing towards a more targeted, power-efficient, form of hardware, the Field-Programmable Gate Array (FPGA). These devices allow the user to programmatically tailor the model processing architecture, resulting in increased inference performance and lower power demands. Their resources however are limited, which leads to the necessity of simplifying the target deep machine learning models. Dynamic Deep Neural Networks (DNNs) are a class of models that go beyond limits of static model compression, by tuning computational workload to the difficultly of inputs on a per-sample basis. In spite of the model simplification capabilities of Dynamic DNNs and the provable efficiency of FPGAs, little work has been done towards accelerating Dynamic DNNs on FPGAs. In this paper we discuss why this occurs by highlighting the challenges and limitations, both at the software and hardware level. We detail the available efficiency, performance gains, and practical benefits of state-of-the-art Dynamic DNN implementations when FPGAs are adopted as the acceleration device. Finally, we present our conclusions and recommendations for continued research in this space

Text
FPGA Acceleration of Dynamic Neural Networks Challenges and Advancements - Accepted Manuscript
Available under License Creative Commons Attribution.
Download (420kB)

More information

Accepted/In Press date: 29 July 2024
e-pub ahead of print date: 15 August 2024
Venue - Dates: IEEE International Conference on Omni-layer Intelligent Systems, King's College, London, United Kingdom, 2024-07-29 - 2024-07-31

Identifiers

Local EPrints ID: 499524
URI: http://eprints.soton.ac.uk/id/eprint/499524
PURE UUID: cc11a383-5042-4104-b6ae-ee0818db9516
ORCID for Jonathon Hare: ORCID iD orcid.org/0000-0003-2921-4283
ORCID for Geoff V. Merrett: ORCID iD orcid.org/0000-0003-4980-3894

Catalogue record

Date deposited: 24 Mar 2025 17:59
Last modified: 25 Mar 2025 05:01

Export record

Altmetrics

Contributors

Author: Anastasios Dimitriou
Author: Benjamin Biggs
Author: Jonathon Hare ORCID iD
Author: Geoff V. Merrett ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

Loading...

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

Back to top

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×