Text-driven automatic image sequence generation using facial modeling for digital TV news production system

Research output: Journal Publications and Reviews (RGC: 21, 22, 62)22_Publication in policy or professional journal

View graph of relations

Author(s)

Related Research Unit(s)

Detail(s)

Original languageEnglish
Pages (from-to)1409-1412
Journal / PublicationProceedings - IEEE International Symposium on Circuits and Systems
Volume2
Publication statusPublished - 1997

Conference

Title1997 IEEE International Symposium on Circuits and Systems (ISCAS '97)
PlaceChina
CityHong Kong
Period9 - 12 June 1997

Abstract

This paper presents a facial modeling approach for automating the head-and-shoulder image sequence generation for digital TV news video clips production which is the most expensive part in terms of manpower and cost. With the IPA phonetics transcribed from news script being the driving parameter, the high precise adapted 2-D wireframe model on a frontal view of speaker image with sufficient facial textural information will be used to define the associated facial action units (AFAUs) corresponding to the phonemes. This developed facial modeling approach increases the intelligibility of facial non-verbal communication for potential audiovisual lip-synch application like TV news video clips, video telephony, Story Teller On Demand (STOD), lip-reading for the deaf or the hearing-impaired.