Learning Dynamic Memory Networks for Object Tracking

Tianyu Yang*, Antoni B. Chan

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

98 Citations (Scopus)

Abstract

Template-matching methods for visual tracking have gained popularity recently due to their comparable performance and fast speed. However, they lack effective ways to adapt to changes in the target object’s appearance, making their tracking accuracy still far from state-of-the-art. In this paper, we propose a dynamic memory network to adapt the template to the target’s appearance variations during tracking. An LSTM is used as a memory controller, where the input is the search feature map and the outputs are the control signals for the reading and writing process of the memory block. As the location of the target is at first unknown in the search feature map, an attention mechanism is applied to concentrate the LSTM input on the potential target. To prevent aggressive model adaptivity, we apply gated residual template learning to control the amount of retrieved memory that is used to combine with the initial template. Unlike tracking-by-detection methods where the object’s information is maintained by the weight parameters of neural networks, which requires expensive online fine-tuning to be adaptable, our tracker runs completely feed-forward and adapts to the target’s appearance changes by updating the external memory. Moreover, unlike other tracking methods where the model capacity is fixed after offline training – the capacity of our tracker can be easily enlarged as the memory requirements of a task increase, which is favorable for memorizing long-term object information. Extensive experiments on OTB and VOT demonstrates that our tracker MemTrack performs favorably against state-of-the-art tracking methods while retaining real-time speed of 50 fps.
Original languageEnglish
Title of host publicationComputer Vision – ECCV 2018
Subtitle of host publication15th European Conference, 2018, Proceedings
EditorsVittorio Ferrari, Martial Hebert, Cristian Sminchisescu
PublisherSpringer Verlag
Pages153-169
ISBN (Electronic)9783030012403
ISBN (Print)9783030012397
DOIs
Publication statusPublished - Sept 2018
Event15th European Conference on Computer Vision (ECCV 2018) - Munich, Germany
Duration: 8 Sept 201814 Sept 2018
http://openaccess.thecvf.com/content_ECCV_2018/html/Tianyu_Yang_Learning_Dynamic_Memory_ECCV_2018_paper.html
https://eccv2018.org/
https://eccv2018.org/wp-content/uploads/2018/09/ECCV_2018_final.pdf
https://sites.google.com/view/eccvfashion/artworks

Publication series

NameLecture Notes in Computer Science
Volume11213
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference15th European Conference on Computer Vision (ECCV 2018)
Abbreviated titleECCV 2018
Country/TerritoryGermany
CityMunich
Period8/09/1814/09/18
Internet address

Bibliographical note

Research Unit(s) information for this publication is provided by the author(s) concerned.

Research Keywords

  • Addressable memory
  • Gated residual template learning

Fingerprint

Dive into the research topics of 'Learning Dynamic Memory Networks for Object Tracking'. Together they form a unique fingerprint.

Cite this