A Hierarchical Learning Approach for Human Action Recognition

Sensors (Basel). 2020 Sep 1;20(17):4946. doi: 10.3390/s20174946.

Abstract

In the domain of human action recognition, existing works mainly focus on using RGB, depth, skeleton and infrared data for analysis. While these methods have the benefit of being non-invasive, they can only be used within limited setups, are prone to issues such as occlusion and often need substantial computational resources. In this work, we address human action recognition through inertial sensor signals, which have a vast quantity of practical applications in fields such as sports analysis and human-machine interfaces. For that purpose, we propose a new learning framework built around a 1D-CNN architecture, which we validated by achieving very competitive results on the publicly available UTD-MHAD dataset. Moreover, the proposed method provides some answers to two of the greatest challenges currently faced by action recognition algorithms, which are (1) the recognition of high-level activities and (2) the reduction of their computational cost in order to make them accessible to embedded devices. Finally, this paper also investigates the tractability of the features throughout the proposed framework, both in time and duration, as we believe it could play an important role in future works in order to make the solution more intelligible, hardware-friendly and accurate.

Keywords: 1D CNN; HAR; explainable deep-learning; high-level activity recognition; human action recognition; intelligent sensors; light-weight method; modular method; one-vs.-all.

Publication types

  • Letter

MeSH terms

  • Algorithms
  • Deep Learning*
  • Human Activities*
  • Humans
  • Skeleton
  • Sports*