Comparing Synchronicity in Body Movement among Jazz Musicians with Their Emotions

Sensors (Basel). 2023 Jul 29;23(15):6789. doi: 10.3390/s23156789.

Abstract

This paper presents novel preliminary research that investigates the relationship between the flow of a group of jazz musicians, quantified through multi-person pose synchronization, and their collective emotions. We have developed a real-time software to calculate the physical synchronicity of team members by tracking the difference in arm, leg, and head movements using Lightweight OpenPose. We employ facial expression recognition to evaluate the musicians' collective emotions. Through correlation and regression analysis, we establish that higher levels of synchronized body and head movements correspond to lower levels of disgust, anger, sadness, and higher levels of joy among the musicians. Furthermore, we utilize 1-D CNNs to predict the collective emotions of the musicians. The model leverages 17 body synchrony keypoint vectors as features, resulting in a training accuracy of 61.47% and a test accuracy of 66.17%.

Keywords: affective computing; collective behaviour analysis; convolutional neural networks; facial emotion recognition; multi-person pose synchronization; pose estimation.

Publication types

  • Comparative Study

MeSH terms

  • Disgust*
  • Emotions
  • Facial Expression
  • Facial Recognition*
  • Head Movements
  • Humans