Deep-learning-based unobtrusive handedness prediction for one-handed smartphone interaction

Taizhou Chen, Kening Zhu*, Ming Chieh Yang

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

3 Citations (Scopus)

Abstract

The handedness (i.e. the side of the holding and operating hand) is an important contextual information to optimise the one-handed smartphone interaction. In this paper, we present a deep-learning-based technique for unobtrusive handedness prediction in one-handed smartphone interaction. Our approach is built upon a multilayer LSTM (Long-Short-Term Memory) neural network, and processes the built-in motion-sensor data of the phone in real time. Compared to the existing approaches, our approach eliminates the need of extra user actions (e.g., on-screen tapping and swiping), and predicts the handedness based on the picking-up action and the holding posture before the user performs any operation on the screen. Our approach is able to predict the handedness when a user is sitting, standing, and walking at an accuracy of 97.4%, 94.6%, and 92.4%, respectively. We also show that our approach is robust to the turbulent noise with an average accuracy of 94.6% for the situations of users in the transportation tools (e.g., bus, train, and scooter). Furthermore, the presented approach can classify users’ real-life single-handed smartphone usage into left- and right-handed with an average accuracy of 89.2%.
Original languageEnglish
Pages (from-to)4941-4964
JournalMultimedia Tools and Applications
Volume82
Issue number4
Online published18 Jan 2022
DOIs
Publication statusPublished - Feb 2023

Research Keywords

  • Handedness prediction
  • LSTM
  • Motion sensor
  • Single hand
  • Smartphone interaction

Fingerprint

Dive into the research topics of 'Deep-learning-based unobtrusive handedness prediction for one-handed smartphone interaction'. Together they form a unique fingerprint.

Cite this