Summarizing rushes videos by motion, object, and event understanding

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

48 Scopus Citations
View graph of relations

Author(s)

Related Research Unit(s)

Detail(s)

Original languageEnglish
Article number5993544
Pages (from-to)76-87
Journal / PublicationIEEE Transactions on Multimedia
Volume14
Issue number1
Publication statusPublished - Feb 2012

Abstract

Rushes footages are considered as cheap gold mine with the potential for reuse in broadcasting and filmmaking industries. However, mining gold from unedited videos such as rushes is challenging as the reusable segments are buried in a large set of redundant information. In this paper, we propose a unified framework for stock footage classification and summarization to support video editors in navigating and organizing rushes videos. Our approach is composed of two steps. First, we employ motion features to filter the undesired camera motion and locate the stock footage. A hierarchical hidden Markov model (HHMM) is proposed to model the motion feature distribution and classify video segments into different categories to decide their potential for reuse. Second, we generate a short video summary to facilitate quick browsing of the stock footages by including the objects and events that are important for storytelling. For objects, we detect the presence of persons and moving objects. For events, we extract a set of features to detect and describe visual (motion activities and scene changes) and audio events (speech clips). A representability measure is then proposed to select the most representative video clips for video summarization. Our experiments show that the proposed HHMM significantly outperforms other methods based on SVM, FSM, and HMM. The automatically generated rushes summaries are also demonstrated to be easy-to-understand, containing little redundancy, and capable of including ground-truth objects and events with shorter durations and relatively pleasant rhythm based on the TRECVID 2007, 2008, and our subjective evaluations. © 2006 IEEE.

Research Area(s)

  • Motion analysis, object and event understanding, rushes video structuring, video summarization