Multi-Modal Residual Perceptron Network for Audio-Video Emotion Recognition

Sensors (Basel). 2021 Aug 12;21(16):5452. doi: 10.3390/s21165452.

Abstract

Emotion recognition is an important research field for human-computer interaction. Audio-video emotion recognition is now attacked with deep neural network modeling tools. In published papers, as a rule, the authors show only cases of the superiority in multi-modality over audio-only or video-only modality. However, there are cases of superiority in uni-modality that can be found. In our research, we hypothesize that for fuzzy categories of emotional events, the within-modal and inter-modal noisy information represented indirectly in the parameters of the modeling neural network impedes better performance in the existing late fusion and end-to-end multi-modal network training strategies. To take advantage of and overcome the deficiencies in both solutions, we define a multi-modal residual perceptron network which performs end-to-end learning from multi-modal network branches, generalizing better multi-modal feature representation. For the proposed multi-modal residual perceptron network and the novel time augmentation for streaming digital movies, the state-of-the-art average recognition rate was improved to 91.4% for the Ryerson Audio-Visual Database of Emotional Speech and Song dataset and to 83.15% for the Crowd-Sourced Emotional Multi Modal Actors dataset. Moreover, the multi-modal residual perceptron network concept shows its potential for multi-modal applications dealing with signal sources not only of optical and acoustical types.

Keywords: audio sensor; deep features fusion; deep neural network; emotion recognition; multi-modal classifier; video sensor.

MeSH terms

  • Emotions*
  • Humans
  • Motion Pictures
  • Neural Networks, Computer*
  • Speech