SOWP: Spatially ordered and weighted patch descriptor for visual tracking

Han Ul Kim, Dae Youn Lee, Jae Young Sim, Chang-Su Kim

Research output: Chapter in Book/Report/Conference proceedingConference contribution

88 Citations (Scopus)

Abstract

A simple yet effective object descriptor for visual tracking is proposed in this paper. We first decompose the bounding box of a target object into multiple patches, which are described by color and gradient histograms. Then, we concatenate the features of the spatially ordered patches to represent the object appearance. Moreover, to alleviate the impacts of background information possibly included in the bounding box, we determine patch weights using random walk with restart (RWR) simulations. The patch weights represent the importance of each patch in the description of foreground information, and are used to construct an object descriptor, called spatially ordered and weighted patch (SOWP) descriptor. We incorporate the proposed SOWP descriptor into the structured output tracking framework. Experimental results demonstrate that the proposed algorithm yields significantly better performance than the state-of-the-art trackers on a recent benchmark dataset, and also excels in another recent benchmark dataset.

Original languageEnglish
Title of host publicationProceedings - 2015 IEEE International Conference on Computer Vision, ICCV 2015
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages3011-3019
Number of pages9
Volume11-18-December-2015
ISBN (Electronic)9781467383912
DOIs
Publication statusPublished - 2016 Feb 17
Event15th IEEE International Conference on Computer Vision, ICCV 2015 - Santiago, Chile
Duration: 2015 Dec 112015 Dec 18

Other

Other15th IEEE International Conference on Computer Vision, ICCV 2015
Country/TerritoryChile
CitySantiago
Period15/12/1115/12/18

ASJC Scopus subject areas

  • Software
  • Computer Vision and Pattern Recognition

Fingerprint

Dive into the research topics of 'SOWP: Spatially ordered and weighted patch descriptor for visual tracking'. Together they form a unique fingerprint.

Cite this