FPGA acceleration of dynamic neural networks: challenges and advancements
FPGA acceleration of dynamic neural networks: challenges and advancements
Modern machine learning methods continue to produce models with a high memory footprint and computational complexity that are increasingly difficult to deploy in resource constrained environments. This is, in part, driven by a focus on costly, power-intensive GPUs, which has a feedback effect on the variety of methods and models chosen for development. We advocate for a transition away from the general purpose processing towards a more targeted, power-efficient, form of hardware, the Field-Programmable Gate Array (FPGA). These devices allow the user to programmatically tailor the model processing architecture, resulting in increased inference performance and lower power demands. Their resources however are limited, which leads to the necessity of simplifying the target deep machine learning models. Dynamic Deep Neural Networks (DNNs) are a class of models that go beyond limits of static model compression, by tuning computational workload to the difficultly of inputs on a per-sample basis. In spite of the model simplification capabilities of Dynamic DNNs and the provable efficiency of FPGAs, little work has been done towards accelerating Dynamic DNNs on FPGAs. In this paper we discuss why this occurs by highlighting the challenges and limitations, both at the software and hardware level. We detail the available efficiency, performance gains, and practical benefits of state-of-the-art Dynamic DNN implementations when FPGAs are adopted as the acceleration device. Finally, we present our conclusions and recommendations for continued research in this space
Dimitriou, Anastasios
02f87799-17dc-4271-96c3-8b30e64e659e
Biggs, Benjamin
8933978b-da66-4f37-bc72-58e41137d940
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
Merrett, Geoff V.
89b3a696-41de-44c3-89aa-b0aa29f54020
Dimitriou, Anastasios
02f87799-17dc-4271-96c3-8b30e64e659e
Biggs, Benjamin
8933978b-da66-4f37-bc72-58e41137d940
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
Merrett, Geoff V.
89b3a696-41de-44c3-89aa-b0aa29f54020
Dimitriou, Anastasios, Biggs, Benjamin, Hare, Jonathon and Merrett, Geoff V.
(2024)
FPGA acceleration of dynamic neural networks: challenges and advancements.
In Proceedings of 2024 IEEE International Conference on Omni-layer Intelligent Systems (COINS).
6 pp
.
(doi:10.1109/COINS61597.2024.10622127).
Record type:
Conference or Workshop Item
(Paper)
Abstract
Modern machine learning methods continue to produce models with a high memory footprint and computational complexity that are increasingly difficult to deploy in resource constrained environments. This is, in part, driven by a focus on costly, power-intensive GPUs, which has a feedback effect on the variety of methods and models chosen for development. We advocate for a transition away from the general purpose processing towards a more targeted, power-efficient, form of hardware, the Field-Programmable Gate Array (FPGA). These devices allow the user to programmatically tailor the model processing architecture, resulting in increased inference performance and lower power demands. Their resources however are limited, which leads to the necessity of simplifying the target deep machine learning models. Dynamic Deep Neural Networks (DNNs) are a class of models that go beyond limits of static model compression, by tuning computational workload to the difficultly of inputs on a per-sample basis. In spite of the model simplification capabilities of Dynamic DNNs and the provable efficiency of FPGAs, little work has been done towards accelerating Dynamic DNNs on FPGAs. In this paper we discuss why this occurs by highlighting the challenges and limitations, both at the software and hardware level. We detail the available efficiency, performance gains, and practical benefits of state-of-the-art Dynamic DNN implementations when FPGAs are adopted as the acceleration device. Finally, we present our conclusions and recommendations for continued research in this space
Text
FPGA Acceleration of Dynamic Neural Networks Challenges and Advancements
- Accepted Manuscript
More information
Accepted/In Press date: 29 July 2024
e-pub ahead of print date: 15 August 2024
Venue - Dates:
IEEE International Conference on Omni-layer Intelligent Systems, King's College, London, United Kingdom, 2024-07-29 - 2024-07-31
Identifiers
Local EPrints ID: 499524
URI: http://eprints.soton.ac.uk/id/eprint/499524
PURE UUID: cc11a383-5042-4104-b6ae-ee0818db9516
Catalogue record
Date deposited: 24 Mar 2025 17:59
Last modified: 25 Mar 2025 05:01
Export record
Altmetrics
Contributors
Author:
Anastasios Dimitriou
Author:
Benjamin Biggs
Author:
Jonathon Hare
Author:
Geoff V. Merrett
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
Loading...
View more statistics