GestOnHMD: Enabling Gesture-based Interaction on Low-cost VR Head-Mounted Display

IEEE Trans Vis Comput Graph. 2021 May;27(5):2597-2607. doi: 10.1109/TVCG.2021.3067689. Epub 2021 Apr 15.

Abstract

Low-cost virtual-reality (VR) head-mounted displays (HMDs) with the integration of smartphones have brought the immersive VR to the masses, and increased the ubiquity of VR. However, these systems are often limited by their poor interactivity. In this paper, we present GestOnHMD, a gesture-based interaction technique and a gesture-classification pipeline that leverages the stereo microphones in a commodity smartphone to detect the tapping and the scratching gestures on the front, the left, and the right surfaces on a mobile VR headset. Taking the Google Cardboard as our focused headset, we first conducted a gesture-elicitation study to generate 150 user-defined gestures with 50 on each surface. We then selected 15, 9, and 9 gestures for the front, the left, and the right surfaces respectively based on user preferences and signal detectability. We constructed a data set containing the acoustic signals of 18 users performing these on-surface gestures, and trained the deep-learning classification pipeline for gesture detection and recognition. Lastly, with the real-time demonstration of GestOnHMD, we conducted a series of online participatory-design sessions to collect a set of user-defined gesture-referent mappings that could potentially benefit from GestOnHMD.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Acoustics
  • Adult
  • Computer Graphics / instrumentation*
  • Deep Learning
  • Equipment Design
  • Female
  • Gestures*
  • Hand / physiology
  • Humans
  • Male
  • Signal Processing, Computer-Assisted
  • Smart Glasses*
  • Smartphone
  • Virtual Reality*
  • Young Adult