Efficient Federated Learning: a Tsetlin Machine-based approach
Efficient Federated Learning: a Tsetlin Machine-based approach
Federated Learning (FL) is a paradigm for collaboratively training a global model across distributed client devices, in which training occurs on-device and only model parameters are communicated to a central server, thereby preserving data privacy. However, its adoption in practice is constrained by communication overhead, computational limitations, and statistical data heterogeneity. Motivated by the logic-based formulation and bit-level representation of Tsetlin Machines (TMs), which enable a low-complexity design, this dissertation addresses FL challenges and proposes a family of TM-based FL frameworks as lightweight, resource-efficient alternatives to conventional deep neural network (DNN)–based FL approaches.The first contribution, FedTM, introduces the first FL framework to capitalise on the bit-level structure and low-complexity learning dynamics of TMs in a federated setting. Building on this, the second contribution, CS-pFedTM, presents a personalised FL framework that addresses statistical heterogeneity through adaptive clause allocation, driven by data heterogeneity and communication constraints. Finally, FedTMOS proposes an efficient one-shot FL framework that eliminates server-side training by clustering local models and constructing a global model in a single communication round via class-adaptive parameter organisation.These methods establish a set of TM-based FL algorithms that are robust to statistical heterogeneity and efficient in terms of communication, computation and memory. Extensive evaluations on image and audio datasets demonstrate competitive performance–resource trade-offs relative to state-of-the-art baselines. While TM research remains at an early stage, this dissertation lays the foundations for TM-based FL and highlights promising directions for future work, including on-device implementations and the exploration of pre-training and transfer learning strategies.
University of Southampton
How, Shannon Shi Qi
f0753707-bb20-42e3-a1aa-bebf9f6c9545
2026
How, Shannon Shi Qi
f0753707-bb20-42e3-a1aa-bebf9f6c9545
Merrett, Geoff
89b3a696-41de-44c3-89aa-b0aa29f54020
Chauhan, Jagmohan
831a12dc-6df9-40ea-8bb3-2c5da8882804
Hare, Jonathon
65ba2cda-eaaf-4767-a325-cd845504e5a9
How, Shannon Shi Qi
(2026)
Efficient Federated Learning: a Tsetlin Machine-based approach.
University of Southampton, Doctoral Thesis, 137pp.
Record type:
Thesis
(Doctoral)
Abstract
Federated Learning (FL) is a paradigm for collaboratively training a global model across distributed client devices, in which training occurs on-device and only model parameters are communicated to a central server, thereby preserving data privacy. However, its adoption in practice is constrained by communication overhead, computational limitations, and statistical data heterogeneity. Motivated by the logic-based formulation and bit-level representation of Tsetlin Machines (TMs), which enable a low-complexity design, this dissertation addresses FL challenges and proposes a family of TM-based FL frameworks as lightweight, resource-efficient alternatives to conventional deep neural network (DNN)–based FL approaches.The first contribution, FedTM, introduces the first FL framework to capitalise on the bit-level structure and low-complexity learning dynamics of TMs in a federated setting. Building on this, the second contribution, CS-pFedTM, presents a personalised FL framework that addresses statistical heterogeneity through adaptive clause allocation, driven by data heterogeneity and communication constraints. Finally, FedTMOS proposes an efficient one-shot FL framework that eliminates server-side training by clustering local models and constructing a global model in a single communication round via class-adaptive parameter organisation.These methods establish a set of TM-based FL algorithms that are robust to statistical heterogeneity and efficient in terms of communication, computation and memory. Extensive evaluations on image and audio datasets demonstrate competitive performance–resource trade-offs relative to state-of-the-art baselines. While TM research remains at an early stage, this dissertation lays the foundations for TM-based FL and highlights promising directions for future work, including on-device implementations and the exploration of pre-training and transfer learning strategies.
Text
Dissertation_ShannonHow_pdfa
- Version of Record
Restricted to Repository staff only until 5 May 2027.
Text
Final-thesis-submission-Examination-Miss-Shannon-How
Restricted to Repository staff only
More information
Published date: 2026
Identifiers
Local EPrints ID: 511247
URI: http://eprints.soton.ac.uk/id/eprint/511247
PURE UUID: ef230f13-ef80-4cfa-a4ea-63a3e52e2b2d
Catalogue record
Date deposited: 08 May 2026 17:06
Last modified: 09 May 2026 02:15
Export record
Altmetrics
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics