Computational validation of the motor contribution to speech perception

Top Cogn Sci. 2014 Jul;6(3):461-75. doi: 10.1111/tops.12095. Epub 2014 Jun 17.

Abstract

Action perception and recognition are core abilities fundamental for human social interaction. A parieto-frontal network (the mirror neuron system) matches visually presented biological motion information onto observers' motor representations. This process of matching the actions of others onto our own sensorimotor repertoire is thought to be important for action recognition, providing a non-mediated "motor perception" based on a bidirectional flow of information along the mirror parieto-frontal circuits. State-of-the-art machine learning strategies for hand action identification have shown better performances when sensorimotor data, as opposed to visual information only, are available during learning. As speech is a particular type of action (with acoustic targets), it is expected to activate a mirror neuron mechanism. Indeed, in speech perception, motor centers have been shown to be causally involved in the discrimination of speech sounds. In this paper, we review recent neurophysiological and machine learning-based studies showing (a) the specific contribution of the motor system to speech perception and (b) that automatic phone recognition is significantly improved when motor data are used during training of classifiers (as opposed to learning from purely auditory data).

Keywords: Automatic speech recognition; Machine learning; Motor theory of speech perception; Transcranial magnetic stimulation.

Publication types

  • Research Support, Non-U.S. Gov't
  • Review

MeSH terms

  • Artificial Intelligence*
  • Humans
  • Pattern Recognition, Automated*
  • Psychomotor Performance / physiology*
  • Recognition, Psychology / physiology
  • Speech Perception / physiology*