Skip to main navigation Skip to search Skip to main content

Prediction-based gesture detection in lecture videos by combining visual, speech and electronic slides

Feng Wang, Chong-Wah Ngo, Ting-Chuen Pong

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

This paper presents an efficient algorithm for gesture detection in lecture videos by combining visual, speech and electronic slides. Besides accuracy, response time is also considered to cope with the efficiency requirements of real-time applications. Candidate gestures are first detected by visual cue. Then we modifity HMM models for complete gestures to predict and recognize incomplete gestures before the whole gestures paths are observed. Gesture recognition is used to verify the results of gesture detection. The relations between visual, speech and slides are analyzed. The correspondence between speech and gesture is employed to improve the accuracy and the responsiveness of gesture detection. © 2006 IEEE.
Original languageEnglish
Title of host publication2006 IEEE International Conference on Multimedia and Expo, ICME 2006 - Proceedings
Pages653-656
Volume2006
DOIs
Publication statusPublished - 2006
Event2006 IEEE International Conference on Multimedia and Expo, ICME 2006 - Toronto, ON, Canada
Duration: 9 Jul 200612 Jul 2006

Publication series

Name
Volume2006

Conference

Conference2006 IEEE International Conference on Multimedia and Expo, ICME 2006
PlaceCanada
CityToronto, ON
Period9/07/0612/07/06

Fingerprint

Dive into the research topics of 'Prediction-based gesture detection in lecture videos by combining visual, speech and electronic slides'. Together they form a unique fingerprint.

Cite this