Sounds of COVID-19: exploring realistic performance of audio-based digital testing
Sounds of COVID-19: exploring realistic performance of audio-based digital testing
To identify Coronavirus disease (COVID-19) cases efficiently, affordably, and at scale, recent work has shown how audio (including cough, breathing and voice) based approaches can be used for testing. However, there is a lack of exploration of how biases and methodological decisions impact these tools’ performance in practice. In this paper, we explore the realistic performance of audio-based digital testing of COVID-19. To investigate this, we collected a large crowdsourced respiratory audio dataset through a mobile app, alongside symptoms and COVID-19 test results. Within the collected dataset, we selected 5240 samples from 2478 English-speaking participants and split them into participant-independent sets for model development and validation. In addition to controlling the language, we also balanced demographics for model training to avoid potential acoustic bias. We used these audio samples to construct an audio-based COVID-19 prediction model. The unbiased model took features extracted from breathing, coughs and voice signals as predictors and yielded an AUC-ROC of 0.71 (95% CI: 0.65–0.77). We further explored several scenarios with different types of unbalanced data distributions to demonstrate how biases and participant splits affect the performance. With these different, but less appropriate, evaluation strategies, the performance could be overestimated, reaching an AUC up to 0.90 (95% CI: 0.85–0.95) in some circumstances. We found that an unrealistic experimental setting can result in misleading, sometimes over-optimistic, performance. Instead, we reported complete and reliable results on crowd-sourced data, which would allow medical professionals and policy makers to accurately assess the value of this technology and facilitate its deployment.
Han, Jing
0e18bcab-5434-4606-b635-70fb07250322
Xia, Tong
763cc71a-eaf1-4902-a27d-62cd5912a569
Spathis, Dimitris
0352ccc2-a97c-4273-998a-0bc8546612fa
Bondareva, Erika
7126e0d3-4e24-4e59-90bc-d56aaa8a639c
Brown, Chloë
b65ca127-9b66-4a6e-9e3b-95c23755732d
Chauhan, Jagmohan
831a12dc-6df9-40ea-8bb3-2c5da8882804
Dang, Ting
1ef75748-8531-43ef-9edd-9553d0899940
Grammenos, Andreas
a0e7ff5f-2149-4aab-b3e3-6733d7290659
Hasthanasombat, Apinan
94b01385-5bf6-455d-a1f9-06a9c3149ba5
Floto, Andres
79ab6e97-aee4-441c-81ba-8d829c057043
Cicuta, Pietro
80bc9499-6c6a-4d7a-8d45-b5b3b76a4695
Mascolo, Cecilia
e4a7bcf7-72c8-43b7-b6b3-4f8980da245d
28 January 2022
Han, Jing
0e18bcab-5434-4606-b635-70fb07250322
Xia, Tong
763cc71a-eaf1-4902-a27d-62cd5912a569
Spathis, Dimitris
0352ccc2-a97c-4273-998a-0bc8546612fa
Bondareva, Erika
7126e0d3-4e24-4e59-90bc-d56aaa8a639c
Brown, Chloë
b65ca127-9b66-4a6e-9e3b-95c23755732d
Chauhan, Jagmohan
831a12dc-6df9-40ea-8bb3-2c5da8882804
Dang, Ting
1ef75748-8531-43ef-9edd-9553d0899940
Grammenos, Andreas
a0e7ff5f-2149-4aab-b3e3-6733d7290659
Hasthanasombat, Apinan
94b01385-5bf6-455d-a1f9-06a9c3149ba5
Floto, Andres
79ab6e97-aee4-441c-81ba-8d829c057043
Cicuta, Pietro
80bc9499-6c6a-4d7a-8d45-b5b3b76a4695
Mascolo, Cecilia
e4a7bcf7-72c8-43b7-b6b3-4f8980da245d
Han, Jing, Xia, Tong, Spathis, Dimitris, Bondareva, Erika, Brown, Chloë, Chauhan, Jagmohan, Dang, Ting, Grammenos, Andreas, Hasthanasombat, Apinan, Floto, Andres, Cicuta, Pietro and Mascolo, Cecilia
(2022)
Sounds of COVID-19: exploring realistic performance of audio-based digital testing.
npj Digital Medicine, 5 (1), [16].
(doi:10.1038/s41746-021-00553-x).
Abstract
To identify Coronavirus disease (COVID-19) cases efficiently, affordably, and at scale, recent work has shown how audio (including cough, breathing and voice) based approaches can be used for testing. However, there is a lack of exploration of how biases and methodological decisions impact these tools’ performance in practice. In this paper, we explore the realistic performance of audio-based digital testing of COVID-19. To investigate this, we collected a large crowdsourced respiratory audio dataset through a mobile app, alongside symptoms and COVID-19 test results. Within the collected dataset, we selected 5240 samples from 2478 English-speaking participants and split them into participant-independent sets for model development and validation. In addition to controlling the language, we also balanced demographics for model training to avoid potential acoustic bias. We used these audio samples to construct an audio-based COVID-19 prediction model. The unbiased model took features extracted from breathing, coughs and voice signals as predictors and yielded an AUC-ROC of 0.71 (95% CI: 0.65–0.77). We further explored several scenarios with different types of unbalanced data distributions to demonstrate how biases and participant splits affect the performance. With these different, but less appropriate, evaluation strategies, the performance could be overestimated, reaching an AUC up to 0.90 (95% CI: 0.85–0.95) in some circumstances. We found that an unrealistic experimental setting can result in misleading, sometimes over-optimistic, performance. Instead, we reported complete and reliable results on crowd-sourced data, which would allow medical professionals and policy makers to accurately assess the value of this technology and facilitate its deployment.
Text
s41746-021-00553-x
- Version of Record
More information
Accepted/In Press date: 13 December 2021
Published date: 28 January 2022
Identifiers
Local EPrints ID: 490941
URI: http://eprints.soton.ac.uk/id/eprint/490941
ISSN: 2398-6352
PURE UUID: 1dbe16e5-0a63-4dd7-a59d-9ebec35f2ddb
Catalogue record
Date deposited: 10 Jun 2024 16:32
Last modified: 10 Jun 2024 16:34
Export record
Altmetrics
Contributors
Author:
Jing Han
Author:
Tong Xia
Author:
Dimitris Spathis
Author:
Erika Bondareva
Author:
Chloë Brown
Author:
Jagmohan Chauhan
Author:
Ting Dang
Author:
Andreas Grammenos
Author:
Apinan Hasthanasombat
Author:
Andres Floto
Author:
Pietro Cicuta
Author:
Cecilia Mascolo
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics