Multimodal integration of micro-Doppler sonar and auditory signals for behavior classification with convolutional networks

Int J Neural Syst. 2013 Oct;23(5):1350021. doi: 10.1142/S0129065713500214. Epub 2013 Jul 23.

Abstract

The ability to recognize the behavior of individuals is of great interest in the general field of safety (e.g. building security, crowd control, transport analysis, independent living for the elderly). Here we report a new real-time acoustic system for human action and behavior recognition that integrates passive audio and active micro-Doppler sonar signatures over multiple time scales. The system architecture is based on a six-layer convolutional neural network, trained and evaluated using a dataset of 10 subjects performing seven different behaviors. Probabilistic combination of system output through time for each modality separately yields 94% (passive audio) and 91% (micro-Doppler sonar) correct behavior classification; probabilistic multimodal integration increases classification performance to 98%. This study supports the efficacy of micro-Doppler sonar systems in characterizing human actions, which can then be efficiently classified using ConvNets. It also demonstrates that the integration of multiple sources of acoustic information can significantly improve the system's performance.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Computer Systems
  • Doppler Effect
  • Humans
  • Neural Networks, Computer*
  • Social Behavior*
  • Sound Spectrography / methods*