TY - JOUR
T1 - GestOnHMD
T2 - Enabling Gesture-based Interaction on Low-cost VR Head-Mounted Display
AU - Chen, Taizhou
AU - Xu, Lantian
AU - Xu, Xianshan
AU - Zhu, Kening
PY - 2021/5
Y1 - 2021/5
N2 - Low-cost virtual-reality (VR) head-mounted displays (HMDs) with the integration of smartphones have brought the immersive
VR to the masses, and increased the ubiquity of VR. However, these systems are often limited by their poor interactivity. In this paper, we
present GestOnHMD, a gesture-based interaction technique and a gesture-classification pipeline that leverages the stereo microphones
in a commodity smartphone to detect the tapping and the scratching gestures on the front, the left, and the right surfaces on a mobile
VR headset. Taking the Google Cardboard as our focused headset, we first conducted a gesture-elicitation study to generate 150
user-defined gestures with 50 on each surface. We then selected 15, 9, and 9 gestures for the front, the left, and the right surfaces
respectively based on user preferences and signal detectability. We constructed a data set containing the acoustic signals of 18 users
performing these on-surface gestures, and trained the deep-learning classification pipeline for gesture detection and recognition. Lastly,
with the real-time demonstration of GestOnHMD, we conducted a series of online participatory-design sessions to collect a set of
user-defined gesture-referent mappings that could potentially benefit from GestOnHMD.
AB - Low-cost virtual-reality (VR) head-mounted displays (HMDs) with the integration of smartphones have brought the immersive
VR to the masses, and increased the ubiquity of VR. However, these systems are often limited by their poor interactivity. In this paper, we
present GestOnHMD, a gesture-based interaction technique and a gesture-classification pipeline that leverages the stereo microphones
in a commodity smartphone to detect the tapping and the scratching gestures on the front, the left, and the right surfaces on a mobile
VR headset. Taking the Google Cardboard as our focused headset, we first conducted a gesture-elicitation study to generate 150
user-defined gestures with 50 on each surface. We then selected 15, 9, and 9 gestures for the front, the left, and the right surfaces
respectively based on user preferences and signal detectability. We constructed a data set containing the acoustic signals of 18 users
performing these on-surface gestures, and trained the deep-learning classification pipeline for gesture detection and recognition. Lastly,
with the real-time demonstration of GestOnHMD, we conducted a series of online participatory-design sessions to collect a set of
user-defined gesture-referent mappings that could potentially benefit from GestOnHMD.
KW - Acoustics
KW - Gesture
KW - Google Cardboard
KW - Headphones
KW - Internet
KW - Microphones
KW - Mobile VR
KW - Pipelines
KW - Sensors
KW - Smart phones
KW - Smartphone
KW - Virtual Reality
UR - http://www.scopus.com/inward/record.url?scp=85103237741&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-85103237741&origin=recordpage
U2 - 10.1109/TVCG.2021.3067689
DO - 10.1109/TVCG.2021.3067689
M3 - RGC 21 - Publication in refereed journal
SN - 1077-2626
VL - 27
SP - 2597
EP - 2607
JO - IEEE Transactions on Visualization and Computer Graphics
JF - IEEE Transactions on Visualization and Computer Graphics
IS - 5
ER -