An investigation into weight fixing networks
An investigation into weight fixing networks
Deep neural networks require vast computational resources, with data movement between memory and processors dominating energy consumption. This thesis presents novel compression techniques that can reduce these costs by minimising the number of unique parameters in neural networks.
We developed Weight Fixing Networks (WFN), which achieve lossless compression using 56x fewer unique weights and 1.9x lower weight-space entropy than state-of-the-art quantisation approaches. However, this revealed a key limitation: treating all weights based solely on value neglects their position-specific importance.
To address this, we introduced Probabilistic Weight Fixing Networks (PWFN), employing Bayesian neural networks to learn position-specific weight uncertainties. This probabilistic framework enables more intelligent compression decisions, reducing a 5-million parameter transformer to just 296 unique values whilst outperforming state-of-the-art quantisation by 1.6% top-1 accuracy on ImageNet.
Our final contribution, Cluster-On-the-Fly PWFN, integrates clustering directly into training, streamlining the compression process. Unexpectedly, these compressed networks also provide well-calibrated uncertainty estimates, enhancing reliability alongside efficiency.
By focusing on whole-network weight reuse and targeting hardware-friendly values, this work demonstrates that neural networks can be compressed far beyond current limits whilst improving both accuracy and uncertainty quantification. These findings have significant implications for deploying efficient and trustworthy AI on resource-constrained devices.
University of Southampton
Subia-Waud, Christopher
1d5426c0-f3ac-4f02-9dd2-83cdc2a8f2fc
2025
Subia-Waud, Christopher
1d5426c0-f3ac-4f02-9dd2-83cdc2a8f2fc
Dasmahapatra, Srinandan
eb5fd76f-4335-4ae9-a88a-20b9e2b3f698
Weddell, Alex
3d8c4d63-19b1-4072-a779-84d487fd6f03
Subia-Waud, Christopher
(2025)
An investigation into weight fixing networks.
University of Southampton, Doctoral Thesis.
Record type:
Thesis
(Doctoral)
Abstract
Deep neural networks require vast computational resources, with data movement between memory and processors dominating energy consumption. This thesis presents novel compression techniques that can reduce these costs by minimising the number of unique parameters in neural networks.
We developed Weight Fixing Networks (WFN), which achieve lossless compression using 56x fewer unique weights and 1.9x lower weight-space entropy than state-of-the-art quantisation approaches. However, this revealed a key limitation: treating all weights based solely on value neglects their position-specific importance.
To address this, we introduced Probabilistic Weight Fixing Networks (PWFN), employing Bayesian neural networks to learn position-specific weight uncertainties. This probabilistic framework enables more intelligent compression decisions, reducing a 5-million parameter transformer to just 296 unique values whilst outperforming state-of-the-art quantisation by 1.6% top-1 accuracy on ImageNet.
Our final contribution, Cluster-On-the-Fly PWFN, integrates clustering directly into training, streamlining the compression process. Unexpectedly, these compressed networks also provide well-calibrated uncertainty estimates, enhancing reliability alongside efficiency.
By focusing on whole-network weight reuse and targeting hardware-friendly values, this work demonstrates that neural networks can be compressed far beyond current limits whilst improving both accuracy and uncertainty quantification. These findings have significant implications for deploying efficient and trustworthy AI on resource-constrained devices.
Text
pda_version_Thesis_Final-1
- Version of Record
Text
Final-thesis-submission-Examination-Mr-Christopher-Subia-Waud
Restricted to Repository staff only
More information
Published date: 2025
Identifiers
Local EPrints ID: 501903
URI: http://eprints.soton.ac.uk/id/eprint/501903
PURE UUID: 9bd6c19f-4984-4759-a381-dbacf5ce46ac
Catalogue record
Date deposited: 11 Jun 2025 18:29
Last modified: 11 Sep 2025 02:14
Export record
Contributors
Author:
Christopher Subia-Waud
Thesis advisor:
Srinandan Dasmahapatra
Thesis advisor:
Alex Weddell
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics