Your browser doesn't support javascript.
loading
Sounds of COVID-19: exploring realistic performance of audio-based digital testing.
Han, Jing; Xia, Tong; Spathis, Dimitris; Bondareva, Erika; Brown, Chloë; Chauhan, Jagmohan; Dang, Ting; Grammenos, Andreas; Hasthanasombat, Apinan; Floto, Andres; Cicuta, Pietro; Mascolo, Cecilia.
Afiliação
  • Han J; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Xia T; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK. tx229@cam.ac.uk.
  • Spathis D; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Bondareva E; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Brown C; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Chauhan J; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Dang T; ECS, University of Southampton, Southampton, UK.
  • Grammenos A; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Hasthanasombat A; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Floto A; Department of Computer Science and Technology, University of Cambridge, Cambridge, UK.
  • Cicuta P; Department of Medicine, University of Cambridge, Cambridge, UK.
  • Mascolo C; Department of Physics, University of Cambridge, Cambridge, UK.
NPJ Digit Med ; 5(1): 16, 2022 Jan 28.
Article em En | MEDLINE | ID: mdl-35091662
ABSTRACT
To identify Coronavirus disease (COVID-19) cases efficiently, affordably, and at scale, recent work has shown how audio (including cough, breathing and voice) based approaches can be used for testing. However, there is a lack of exploration of how biases and methodological decisions impact these tools' performance in practice. In this paper, we explore the realistic performance of audio-based digital testing of COVID-19. To investigate this, we collected a large crowdsourced respiratory audio dataset through a mobile app, alongside symptoms and COVID-19 test results. Within the collected dataset, we selected 5240 samples from 2478 English-speaking participants and split them into participant-independent sets for model development and validation. In addition to controlling the language, we also balanced demographics for model training to avoid potential acoustic bias. We used these audio samples to construct an audio-based COVID-19 prediction model. The unbiased model took features extracted from breathing, coughs and voice signals as predictors and yielded an AUC-ROC of 0.71 (95% CI 0.65-0.77). We further explored several scenarios with different types of unbalanced data distributions to demonstrate how biases and participant splits affect the performance. With these different, but less appropriate, evaluation strategies, the performance could be overestimated, reaching an AUC up to 0.90 (95% CI 0.85-0.95) in some circumstances. We found that an unrealistic experimental setting can result in misleading, sometimes over-optimistic, performance. Instead, we reported complete and reliable results on crowd-sourced data, which would allow medical professionals and policy makers to accurately assess the value of this technology and facilitate its deployment.

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Tipo de estudo: Prognostic_studies Idioma: En Revista: NPJ Digit Med Ano de publicação: 2022 Tipo de documento: Article País de afiliação: Reino Unido

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Tipo de estudo: Prognostic_studies Idioma: En Revista: NPJ Digit Med Ano de publicação: 2022 Tipo de documento: Article País de afiliação: Reino Unido