Conveying facial expressions to blind and visually impaired persons through a wearable vibrotactile device

PLoS One. 2018 Mar 27;13(3):e0194737. doi: 10.1371/journal.pone.0194737. eCollection 2018.

Abstract

In face-to-face social interactions, blind and visually impaired persons (VIPs) lack access to nonverbal cues like facial expressions, body posture, and gestures, which may lead to impaired interpersonal communication. In this study, a wearable sensory substitution device (SSD) consisting of a head mounted camera and a haptic belt was evaluated to determine whether vibrotactile cues around the waist could be used to convey facial expressions to users and whether such a device is desired by VIPs for use in daily living situations. Ten VIPs (mean age: 38.8, SD: 14.4) and 10 sighted persons (SPs) (mean age: 44.5, SD: 19.6) participated in the study, in which validated sets of pictures, silent videos, and videos with audio of facial expressions were presented to the participant. A control measurement was first performed to determine how accurately participants could identify facial expressions while relying on their functional senses. After a short training, participants were asked to determine facial expressions while wearing the emotion feedback system. VIPs using the device showed significant improvements in their ability to determine which facial expressions were shown. A significant increase in accuracy of 44.4% was found across all types of stimuli when comparing the scores of the control (mean±SEM: 35.0±2.5%) and supported (mean±SEM: 79.4±2.1%) phases. The greatest improvements achieved with the support of the SSD were found for silent stimuli (68.3% for pictures and 50.8% for silent videos). SPs also showed consistent, though not statistically significant, improvements while supported. Overall, our study shows that vibrotactile cues are well suited to convey facial expressions to VIPs in real-time. Participants became skilled with the device after a short training session. Further testing and development of the SSD is required to improve its accuracy and aesthetics for potential daily use.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Acoustic Stimulation
  • Adolescent
  • Adult
  • Blindness / pathology
  • Blindness / psychology*
  • Facial Expression*
  • Female
  • Humans
  • Male
  • Middle Aged
  • Photic Stimulation
  • Task Performance and Analysis
  • Visually Impaired Persons / psychology*
  • Wearable Electronic Devices*
  • Young Adult

Grants and funding

The research was funded by: Stichting voor de Technische Wetenschappen (STW): Take Off - Grant number: 15666, http://www.stw.nl/takeoff; Oost NV: INTERREG-project MIND (Medische Innovaties Nederland Duitsland) - Grant number: 122035, http://www.oostnv.nl/page/grensoverschrijdende-medische-innovaties-mind; and The Netherlands Organisation for Health Research and Development (ZonMW): Inzicht - Grant number - 94211004, https://www.zonmw.nl/en/research-and-results/handicap-and-chronicdisease/programmas/programme-detail/insight/. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Tjerk Kostelijk is employed by VicarVision, which developed the FaceReader software used in this study. VicarVision provided support in the form of salary for author TK, but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific role of this author is articulated in the ‘author contributions’ section