American Sign Language Recognition and Translation Using Perception Neuron Wearable Inertial Motion Capture System

Sensors (Basel). 2024 Jan 11;24(2):453. doi: 10.3390/s24020453.

Abstract

Sign language is designed as a natural communication method to convey messages among the deaf community. In the study of sign language recognition through wearable sensors, the data sources are limited, and the data acquisition process is complex. This research aims to collect an American sign language dataset with a wearable inertial motion capture system and realize the recognition and end-to-end translation of sign language sentences with deep learning models. In this work, a dataset consisting of 300 commonly used sentences is gathered from 3 volunteers. In the design of the recognition network, the model mainly consists of three layers: convolutional neural network, bi-directional long short-term memory, and connectionist temporal classification. The model achieves accuracy rates of 99.07% in word-level evaluation and 97.34% in sentence-level evaluation. In the design of the translation network, the encoder-decoder structured model is mainly based on long short-term memory with global attention. The word error rate of end-to-end translation is 16.63%. The proposed method has the potential to recognize more sign language sentences with reliable inertial data from the device.

Keywords: American sign language; deep learning models; wearable inertial sensors.

MeSH terms

  • Humans
  • Motion Capture
  • Neurons
  • Perception
  • Sign Language*
  • United States
  • Wearable Electronic Devices*

Grants and funding

This work was financially supported by Gunma University for the promotion of scientific research.