Hierarchical Encoding of Attended Auditory Objects in Multi-talker Speech Perception

Neuron. 2019 Dec 18;104(6):1195-1209.e3. doi: 10.1016/j.neuron.2019.09.007. Epub 2019 Oct 21.

Abstract

Humans can easily focus on one speaker in a multi-talker acoustic environment, but how different areas of the human auditory cortex (AC) represent the acoustic components of mixed speech is unknown. We obtained invasive recordings from the primary and nonprimary AC in neurosurgical patients as they listened to multi-talker speech. We found that neural sites in the primary AC responded to individual speakers in the mixture and were relatively unchanged by attention. In contrast, neural sites in the nonprimary AC were less discerning of individual speakers but selectively represented the attended speaker. Moreover, the encoding of the attended speaker in the nonprimary AC was invariant to the degree of acoustic overlap with the unattended speaker. Finally, this emergent representation of attended speech in the nonprimary AC was linearly predictable from the primary AC responses. Our results reveal the neural computations underlying the hierarchical formation of auditory objects in human AC during multi-talker speech perception.

Keywords: Heschl’s gyrus; auditory object; cocktail party; encoding; hierarchical; human auditory cortex; multi-talker; speech perception; superior temporal gyrus.

Publication types

  • Research Support, N.I.H., Extramural
  • Research Support, Non-U.S. Gov't

MeSH terms

  • Attention / physiology*
  • Auditory Cortex / physiology*
  • Humans
  • Speech Perception / physiology*