Loading…

Multi-task based object tracking via a collaborative model

•A method for visual object tracking is proposed.•We propose a decomposition model in visual tracking.•The decomposition model is utilized in an integrated generative model and discriminative model tracking framework.•Alternating direction method of multipliers algorithm guarantees the convergence o...

Full description

Saved in:
Bibliographic Details
Published in:Journal of visual communication and image representation 2018-08, Vol.55, p.698-710
Main Authors: Wang, Yong, Luo, Xinbin, Ding, Lu, Hu, Shiqiang
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:•A method for visual object tracking is proposed.•We propose a decomposition model in visual tracking.•The decomposition model is utilized in an integrated generative model and discriminative model tracking framework.•Alternating direction method of multipliers algorithm guarantees the convergence of the decomposition model. This paper presents a multi-task based object tracking algorithm via a collaborative model. Under the framework of particle filtering, we develop a multi-task sparse learning based generative and discriminative classifier model. In the generative model, we propose a histogram based subspace learning method which takes advantage of adaptive templates update. In the discriminative model, we introduce an effective method to compute the confidence value which assigns more weights to the foreground than the background. A decomposition model is employed to take the common features and outliers of each particle into consideration. The alternating direction method of multipliers (ADMM) algorithm guarantees the optimization problem can be solved robustly and accurately. Qualitative and quantitative comparisons with nine state-of-the-art methods demonstrate the effectiveness and efficiency of our method in handling various challenges during tracking.
ISSN:1047-3203
1095-9076
DOI:10.1016/j.jvcir.2018.08.008