Exploration of Decision Sub-Network Architectures for FPGA-based Dynamic DNNs
Exploration of Decision Sub-Network Architectures for FPGA-based Dynamic DNNs
Dynamic Deep Neural Networks (DNNs) can achieve faster execution and less computationally intensive inference by spending fewer resources on easy to recognise or less informative parts of an input. They make data-dependent decisions, which strategically deactivate a model’s components, e.g. layers, channels or sub-networks. However, dynamic DNNs have only been explored and applied on conventional computing systems (CPU+GPU) and programmed with libraries designed for static networks, limiting their effects. In this paper, we propose and explore two approaches for efficiently realising the sub-networks that make these decisions on FPGAs. A pipeline approach targets the use of the existing hardware to execute the sub-network, while a parallel approach uses dedicated circuitry for it. We explore the performance of each using the BranchyNet early exit approach on LeNet-5, and evaluate on a Xilinx ZCU106. The pipeline approach is 36% faster than a desktop CPU. It consumes 0.51 mJ per inference, 16x lower than a non-dynamic network on the same platform and 8x lower than an Nvidia Jetson Xavier NX. The parallel approach executes 17% faster than the pipeline approach when on dynamic inference no early exits are taken, but incurs an increase in energy consumption of 28%.
Dimitriou, Anastasios
02f87799-17dc-4271-96c3-8b30e64e659e
Hu, Mingyu
686551f3-f76b-471d-b424-71a5c68851da
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
Merrett, Geoff
89b3a696-41de-44c3-89aa-b0aa29f54020
19 April 2023
Dimitriou, Anastasios
02f87799-17dc-4271-96c3-8b30e64e659e
Hu, Mingyu
686551f3-f76b-471d-b424-71a5c68851da
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
Merrett, Geoff
89b3a696-41de-44c3-89aa-b0aa29f54020
Dimitriou, Anastasios, Hu, Mingyu, Hare, Jonathon and Merrett, Geoff
(2023)
Exploration of Decision Sub-Network Architectures for FPGA-based Dynamic DNNs.
In Design, Automation and Test in Europe Conference 2023.
2 pp
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
Dynamic Deep Neural Networks (DNNs) can achieve faster execution and less computationally intensive inference by spending fewer resources on easy to recognise or less informative parts of an input. They make data-dependent decisions, which strategically deactivate a model’s components, e.g. layers, channels or sub-networks. However, dynamic DNNs have only been explored and applied on conventional computing systems (CPU+GPU) and programmed with libraries designed for static networks, limiting their effects. In this paper, we propose and explore two approaches for efficiently realising the sub-networks that make these decisions on FPGAs. A pipeline approach targets the use of the existing hardware to execute the sub-network, while a parallel approach uses dedicated circuitry for it. We explore the performance of each using the BranchyNet early exit approach on LeNet-5, and evaluate on a Xilinx ZCU106. The pipeline approach is 36% faster than a desktop CPU. It consumes 0.51 mJ per inference, 16x lower than a non-dynamic network on the same platform and 8x lower than an Nvidia Jetson Xavier NX. The parallel approach executes 17% faster than the pipeline approach when on dynamic inference no early exits are taken, but incurs an increase in energy consumption of 28%.
Text
Exploration of Decision Sub-Network Architectures for FPGA-based Dynamic DNNs
- Accepted Manuscript
More information
Accepted/In Press date: 24 January 2023
Published date: 19 April 2023
Identifiers
Local EPrints ID: 477581
URI: http://eprints.soton.ac.uk/id/eprint/477581
PURE UUID: a6f66ff1-d20d-4c8b-9567-7c00312862c9
Catalogue record
Date deposited: 08 Jun 2023 16:55
Last modified: 17 Mar 2024 03:05
Export record
Contributors
Author:
Anastasios Dimitriou
Author:
Mingyu Hu
Author:
Jonathon Hare
Author:
Geoff Merrett
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics