Department of Computer Science
Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation
© 2014 IEEE. The use of multiple features for tracking has been proved as an effective approach because limitation of each feature could be compensated. Since different types of variations such as illumination, occlusion and pose may happen in a video sequence, especially long sequence videos, how to dynamically select the appropriate features is one of the key problems in this approach. To address this issue in multicue visual tracking, this paper proposes a new joint sparse representation model for robust feature-level fusion. The proposed method dynamically removes unreliable features to be fused for tracking by using the advantages of sparse representation. As a result, robust tracking performance is obtained. Experimental results on publicly available videos show that the proposed method outperforms both existing sparse representation based and fusion-based trackers.
Source Publication Title
Proceedings of 2014 IEEE Conference on Computer Vision and Pattern Recognition
Columbus, United States
Link to Publisher's Edition
Lan, X., Ma, A., & Yuen, P. (2014). Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation. Proceedings of 2014 IEEE Conference on Computer Vision and Pattern Recognition, 1194-1201. https://doi.org/10.1109/CVPR.2014.156