Skip to main navigation Skip to search Skip to main content

In-video product annotation with web information mining

Guangda Li, Meng Wang, Zheng Lu, Richang Hong, Tat-Seng Chua

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

Abstract

Product annotation in videos is of great importance for video browsing, search, and advertisement. However, most of the existing automatic video annotation research focuses on the annotation of high-level concepts, such as events, scenes, and object categories. This article presents a novel solution to the annotation of specific products in videos by mining information from the Web. It collects a set of high-quality training data for each product by simultaneously leveraging Amazon and Google image search engine. A visual signature for each product is then built based on the bag-of-visual-words representation of the training images. A correlative sparsification approach is employed to remove noisy bins in the visual signatures. These signatures are used to annotate video frames. We conduct experiments on more than 1,000 videos and the results demonstrate the feasibility and effectiveness of our approach.
Original languageEnglish
Article number2379797
JournalACM Transactions on Multimedia Computing, Communications and Applications
Volume8
Issue number4
DOIs
Publication statusPublished - 2012
Externally publishedYes

Research Keywords

  • Product Annotation
  • Video search
  • Web mining

Fingerprint

Dive into the research topics of 'In-video product annotation with web information mining'. Together they form a unique fingerprint.

Cite this