The University of Southampton
University of Southampton Institutional Repository

Speech is silver, silence is golden: What do ASVspoof-trained models really learn?

Speech is silver, silence is golden: What do ASVspoof-trained models really learn?
Speech is silver, silence is golden: What do ASVspoof-trained models really learn?
We present our analysis of a significant data artifact in the official 2019/2021 ASVspoof Challenge Dataset. We identify an uneven distribution of silence duration in the training and test splits, which tends to correlate with the target prediction label. Bonafide instances tend to have significantly longer leading and trailing silences than spoofed instances. In this paper, we explore this phenomenon and its impact in depth. We compare several types of models trained on a) only the duration of the leading silence and b) only on the duration of leading and trailing silence. Results show that models trained on only the duration of the leading silence perform particularly well, and achieve up to 85% percent accuracy and an equal error rate (EER) of 15.1%. At the same time, we observe that trimming silence during pre-processing and then training established antispoofing models using signal-based features leads to comparatively worse performance. In that case, EER increases from 3.6% (with silence) to 15.5% (trimmed silence). Our findings suggest that previous work may, in part, have inadvertently learned thespoof/bonafide distinction by relying on the duration of silence as it appears in the official challenge dataset. We discuss the potential consequences that this has for interpreting system scores in the challenge and discuss how the ASV community may further consider this issue.
cs.SD, eess.AS
Müller, Nicolas M.
a318327f-8d2f-4e99-9fff-f7ee0ed7523d
Dieckmann, Franziska
a7590229-80c9-4a45-9afa-7330192b499a
Czempin, Pavel
0a1258a0-8098-4d31-92f7-dec48df28a46
Canals, Roman
cfd99bc0-7fdb-4cbc-a15c-37c1e0644ddb
Böttinger, Konstantin
3f42aacc-85c7-4630-82ba-068ff348b2df
Williams, Jennifer
3a1568b4-8a0b-41d2-8635-14fe69fbb360
Müller, Nicolas M.
a318327f-8d2f-4e99-9fff-f7ee0ed7523d
Dieckmann, Franziska
a7590229-80c9-4a45-9afa-7330192b499a
Czempin, Pavel
0a1258a0-8098-4d31-92f7-dec48df28a46
Canals, Roman
cfd99bc0-7fdb-4cbc-a15c-37c1e0644ddb
Böttinger, Konstantin
3f42aacc-85c7-4630-82ba-068ff348b2df
Williams, Jennifer
3a1568b4-8a0b-41d2-8635-14fe69fbb360

Müller, Nicolas M., Dieckmann, Franziska, Czempin, Pavel, Canals, Roman, Böttinger, Konstantin and Williams, Jennifer (2021) Speech is silver, silence is golden: What do ASVspoof-trained models really learn? ASVspoof 2021 Workshop: An official Interspeech 2021 satellite event,, Online. 16 Sep 2021.

Record type: Conference or Workshop Item (Paper)

Abstract

We present our analysis of a significant data artifact in the official 2019/2021 ASVspoof Challenge Dataset. We identify an uneven distribution of silence duration in the training and test splits, which tends to correlate with the target prediction label. Bonafide instances tend to have significantly longer leading and trailing silences than spoofed instances. In this paper, we explore this phenomenon and its impact in depth. We compare several types of models trained on a) only the duration of the leading silence and b) only on the duration of leading and trailing silence. Results show that models trained on only the duration of the leading silence perform particularly well, and achieve up to 85% percent accuracy and an equal error rate (EER) of 15.1%. At the same time, we observe that trimming silence during pre-processing and then training established antispoofing models using signal-based features leads to comparatively worse performance. In that case, EER increases from 3.6% (with silence) to 15.5% (trimmed silence). Our findings suggest that previous work may, in part, have inadvertently learned thespoof/bonafide distinction by relying on the duration of silence as it appears in the official challenge dataset. We discuss the potential consequences that this has for interpreting system scores in the challenge and discuss how the ASV community may further consider this issue.

Text
2106.12914v4
Download (229kB)

More information

Published date: 23 June 2021
Venue - Dates: ASVspoof 2021 Workshop: An official Interspeech 2021 satellite event,, Online, 2021-09-16 - 2021-09-16
Keywords: cs.SD, eess.AS

Identifiers

Local EPrints ID: 467425
URI: http://eprints.soton.ac.uk/id/eprint/467425
PURE UUID: cd04b8cf-390d-406e-a1ca-93e0d66e2c9d
ORCID for Jennifer Williams: ORCID iD orcid.org/0000-0003-1410-0427

Catalogue record

Date deposited: 08 Jul 2022 16:33
Last modified: 17 Mar 2024 04:12

Export record

Contributors

Author: Nicolas M. Müller
Author: Franziska Dieckmann
Author: Pavel Czempin
Author: Roman Canals
Author: Konstantin Böttinger
Author: Jennifer Williams ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×