The University of Southampton
University of Southampton Institutional Repository

Attacker attribution of audio deepfakes

Attacker attribution of audio deepfakes
Attacker attribution of audio deepfakes
Deepfakes are synthetically generated media often devised with malicious intent. They have become increasingly more convincing with large training datasets advanced neural networks. These fakes are readily being misused for slander, misinformation and fraud. For this reason, intensive research for developing countermeasures is also expanding. However, recent work is almost exclusively limited to deepfake detection - predicting if audio is real or fake. This is despite the fact that attribution (who created which fake?) is an essential building block of a larger defense strategy, as practiced in the field of cybersecurity for a long time. This paper considers the problem of deepfake attacker attribution in the domain of audio. We present several methods for creating attacker signatures using low-level acoustic descriptors and machine learning embeddings. We show that speech signal features are inadequate for characterizing attacker signatures. However, we also demonstrate that embeddings from a recurrent neural network can successfully characterize attacks from both known and unknown attackers. Our attack signature embeddings result in distinct clusters, both for seen and unseen audio deepfakes. We show that these embeddings can be used in downstream-tasks to high-effect, scoring 97.10% accuracy in attacker-id classification.
2958-1796
2788-2792
Williams, Jennifer
3a1568b4-8a0b-41d2-8635-14fe69fbb360
Williams, Jennifer
3a1568b4-8a0b-41d2-8635-14fe69fbb360

Williams, Jennifer (2022) Attacker attribution of audio deepfakes. In Interspeech 2022. pp. 2788-2792 . (doi:10.21437/Interspeech.2022-129).

Record type: Conference or Workshop Item (Paper)

Abstract

Deepfakes are synthetically generated media often devised with malicious intent. They have become increasingly more convincing with large training datasets advanced neural networks. These fakes are readily being misused for slander, misinformation and fraud. For this reason, intensive research for developing countermeasures is also expanding. However, recent work is almost exclusively limited to deepfake detection - predicting if audio is real or fake. This is despite the fact that attribution (who created which fake?) is an essential building block of a larger defense strategy, as practiced in the field of cybersecurity for a long time. This paper considers the problem of deepfake attacker attribution in the domain of audio. We present several methods for creating attacker signatures using low-level acoustic descriptors and machine learning embeddings. We show that speech signal features are inadequate for characterizing attacker signatures. However, we also demonstrate that embeddings from a recurrent neural network can successfully characterize attacks from both known and unknown attackers. Our attack signature embeddings result in distinct clusters, both for seen and unseen audio deepfakes. We show that these embeddings can be used in downstream-tasks to high-effect, scoring 97.10% accuracy in attacker-id classification.

Text
muller22b_interspeech - Version of Record
Restricted to Repository staff only
Request a copy

More information

Published date: 18 September 2022
Venue - Dates: Interspeech 2022, Interspeech, Incheon, Korea, Republic of, 2022-09-18 - 2022-09-22

Identifiers

Local EPrints ID: 501848
URI: http://eprints.soton.ac.uk/id/eprint/501848
ISSN: 2958-1796
PURE UUID: 6fb978e2-5f24-47e5-8eed-298d048c9382
ORCID for Jennifer Williams: ORCID iD orcid.org/0000-0003-1410-0427

Catalogue record

Date deposited: 11 Jun 2025 16:31
Last modified: 22 Aug 2025 02:34

Export record

Altmetrics

Contributors

Author: Jennifer Williams ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×