Loading…

Visual Object Tracking Based on Combination of Local Description and Global Representation

This paper provides a novel method for visual object tracking based on the combination of local scale-invariant feature transform (SIFT) description and global incremental principal component analysis (PCA) representation in loosely constrained conditions. The state of object is defined by the posit...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on circuits and systems for video technology 2011-04, Vol.21 (4), p.408-420
Main Authors: Sun, Li, Liu, Guizhong
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This paper provides a novel method for visual object tracking based on the combination of local scale-invariant feature transform (SIFT) description and global incremental principal component analysis (PCA) representation in loosely constrained conditions. The state of object is defined by the position and shape of a parallelogram, which means that tracking results are given by locating the object in every frame using parallelograms. The whole method is constructed in the framework of particle filter which includes two models: the dynamic model and the observation model. In the dynamic model, particle states are predicted with the help of local SIFT descriptors. Local key point matching between successive frames based on SIFT descriptors provides us an important cue for the prediction of particle states; thus, we can efficiently spread particles in the neighborhood of the predicted position. In the observation model, every particle is evaluated by local key point-weighted incremental PCA representation, which can describe the object more accurately by giving large weights to the pixels in the influence area of key points. Moreover, by incorporating the dynamic forgetting factor, we can update the PCA eigenvectors online according to the object states, which makes our method more adaptable under different situations. Experimental results show that compared to other state-of-the-art methods, the proposed method is robust especially under some difficult conditions, such as strong motion of both object and background, large pose change, and illumination change.
ISSN:1051-8215
1558-2205
DOI:10.1109/TCSVT.2010.2087815