Skip to main navigation Skip to search Skip to main content

Superpixel Tensor Pooling for Visual Tracking Using Multiple Midlevel Visual Cues Fusion

Chong Wu*, Le Zhang, Jiawang Cao, Hong Yan

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

65 Downloads (CityUHK Scholars)

Abstract

In this paper, we propose a method called superpixel tensor pooling tracker which can fuse multiple midlevel cues captured by superpixels into sparse pooled tensor features. Our method first adopts the superpixel method to generate different patches (superpixels) from the target template or candidates. Then for each superpixel, it encodes different midlevel cues including HSI color, RGB color, and spatial coordinates into a histogram matrix to construct a new feature space. Next, these matrices are formed to a third order tensor. After that, the tensor is pooled into the sparse representation. Then the incremental positive and negative subspaces learning is performed. Our method has both good characteristics of midlevel cues and sparse representation hence is more robust to large appearance variations and can capture compact and informative appearance of the target object. To validate the proposed method, we compare it with state-of-the-art methods on 24 sequences with multiple visual tracking challenges. Experiment results demonstrate that our method outperforms them significantly.
Original languageEnglish
Article number8865033
Pages (from-to)147462-147469
Number of pages8
JournalIEEE Access
Volume7
Online published11 Oct 2019
DOIs
Publication statusPublished - 2019

Research Keywords

  • Incremental positive and negative subspaces learning
  • multiple midlevel visual cues fusion
  • superpixel tensor pooling
  • visual tracking

Publisher's Copyright Statement

  • This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/

Fingerprint

Dive into the research topics of 'Superpixel Tensor Pooling for Visual Tracking Using Multiple Midlevel Visual Cues Fusion'. Together they form a unique fingerprint.

Cite this