Detection of COVID-19 from voice, cough and breathing patterns: Dataset and preliminary results

Comput Biol Med. 2021 Nov:138:104944. doi: 10.1016/j.compbiomed.2021.104944. Epub 2021 Oct 13.

Abstract

COVID-19 heavily affects breathing and voice and causes symptoms that make patients' voices distinctive, creating recognizable audio signatures. Initial studies have already suggested the potential of using voice as a screening solution. In this article we present a dataset of voice, cough and breathing audio recordings collected from individuals infected by SARS-CoV-2 virus, as well as non-infected subjects via large scale crowdsourced campaign. We describe preliminary results for detection of COVID-19 from cough patterns using standard acoustic features sets, wavelet scattering features and deep audio embeddings extracted from low-level feature representations (VGGish and OpenL3). Our models achieve accuracy of 88.52%, sensitivity of 88.75% and specificity of 90.87%, confirming the applicability of audio signatures to identify COVID-19 symptoms. We furthermore provide an in-depth analysis of the most informative acoustic features and try to elucidate the mechanisms that alter the acoustic characteristics of coughs of people with COVID-19.

Keywords: Artificial intelligence; COVID-19; Cough; Digital biomarker; Voice.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • COVID-19*
  • Cough / diagnosis
  • Humans
  • Respiration
  • SARS-CoV-2
  • Voice*