A transfer learning framework for predicting the emotional content of generalized sound events

J Acoust Soc Am. 2017 Mar;141(3):1694. doi: 10.1121/1.4977749.

Abstract

Predicting the emotions evoked by generalized sound events is a relatively recent research domain which still needs attention. In this work a framework aiming to reveal potential similarities existing during the perception of emotions evoked by sound events and songs is presented. To this end the following are proposed: (a) the usage of temporal modulation features, (b) a transfer learning module based on an echo state network, and (c) a k-medoids clustering algorithm predicting valence and arousal measurements associated with generalized sound events. The effectiveness of the proposed solution is demonstrated after a thoroughly designed experimental phase employing both sound and music data. The results demonstrate the importance of transfer learning in the specific field and encourage further research on approaches which manage the problem in a synergistic way.

MeSH terms

  • Acoustic Stimulation
  • Adult
  • Algorithms
  • Auditory Perception*
  • Cues
  • Emotions*
  • Female
  • Humans
  • Male
  • Models, Theoretical
  • Music*
  • Pattern Recognition, Physiological
  • Sound*
  • Time Factors
  • Transfer, Psychology*
  • Young Adult