Robust visual tracking via two-stage binocular sparse learning
- Author(s): Ziang Ma 1 ; Wei Lu 1 ; Jun Yin 1 ; Xingming Zhang 1
-
-
View affiliations
-
Affiliations:
1:
Zhejiang Dahua Technology CO., LTD. , Zhejiang Province , Hangzhou , People's Republic of China
-
Affiliations:
1:
Zhejiang Dahua Technology CO., LTD. , Zhejiang Province , Hangzhou , People's Republic of China
- Source:
Volume 2018, Issue 16,
November
2018,
p.
1606 – 1611
DOI: 10.1049/joe.2018.8328 , Online ISSN 2051-3305
Combining multiple features and enforcing joint sparsity have proven to be beneficial for robust tracking. In this study, a novel stereo vision and two-stage sparse representation-based method is presented. First, the colouring information-based features are augmented with a depth view in the appearance modelling of a target object. Unreliable features are then dynamically removed for robust feature-level fusion in the first stage of sparse optimisation. Next, the low rank constraint is imposed onto the objective function, which facilitates a more robust representation of the ensemble of particles over the pruned views. Finally, the authors propose to detect occlusion via depth-based histogram analysis to guarantee the effectiveness of the template update. Experiments are performed on two large-scale benchmark datasets: KITTI and Princeton. Authors’ approach achieves state-of-the-art results in the aspect of robustness and accuracy.
Inspec keywords: object tracking; image fusion; learning (artificial intelligence); feature extraction; image representation; stereo image processing
Other keywords: two-stage sparse representation-based method; robustness; enforcing joint sparsity; robust visual tracking; novel stereo vision; robust feature-level fusion; depth view; colouring information-based features; two-stage binocular sparse learning; sparse optimisation; robust representation; robust tracking; pruned views; unreliable features; low rank constraint; target object; multiple features; depth-based histogram analysis; objective function; appearance modelling
Subjects: Computer vision and image processing techniques; Optical, image and video signal processing
References
-
-
1)
-
14. Babenko, B., Yang, M.H., Belongie, S.: ‘Visual tracking with online multiple instance learning’. The IEEE Conf. Computer Vision and Pattern Recognition, Miami, USA, June 2009, pp. 983–990.
-
-
2)
-
21. Zhang, T., Ghanem, B., Liu, S., et al: ‘Robust visual tracking via structured multi-task sparse learning’, Int. J. Comput. Vis., 2013, 101, (2), pp. 367–383.
-
-
3)
-
29. Erbs, F., Barth, A., Franke, U.: ‘Moving vehicle detection by optimal segmentation of the dynamic stixel world’. Intelligent Vehicles Symp. (IV), Baden, Germany, June 2011, pp. 951–956.
-
-
4)
-
1. Yilmaz, A., Javed, O., Shah, M.: ‘Object tracking: a survey’, ACM Comput. Surv., 2006, 38, (4), pp. 43–56.
-
-
5)
-
3. Smeulders, A. W., Chu, D. M., Cucchiara, R., et al: ‘Visual tracking: an experimental survey’, IEEE Trans. Pattern Anal. Mach. Intell., 2014, 36, (7), pp. 1442–1468.
-
-
6)
-
19. Zhang, T., Ghanem, B., Liu, S., et al: ‘Low-rank sparse learning for robust visual tracking’. European Conf. Computer Vision, Florence, Italy, October 2012, pp. 470–484.
-
-
7)
-
10. Comaniciu, D., Ramesh, V., Meer, P.: ‘Kernel-based object tracking’, IEEE Trans. Pattern Anal. Mach. Intell., 2003, 25, (5), pp. 564–577.
-
-
8)
-
26. Chen, Z., Birchfield, S.T.: ‘Person following with a mobile robot using binocular feature-based tracking’. The IEEE/RSJ Int. Conf. Intelligent Robots and Systems, San Diego, CA, USA, October 2007, pp. 815–820.
-
-
9)
-
30. Danescu, R., Oniga, F., Nedevschi, S.: ‘Modeling and tracking the driving environment with a particle based occupancy grid’, IEEE Trans. Intell. Transp. Syst., 2011, 12, (4), pp. 1331–1342.
-
-
10)
-
28. Bota, S., Nedevschi, S.: ‘Tracking multiple objects in urban traffic environments using dense stereo and optical flow’. The IEEE Conf. Intelligent Transportation Systems (ITSC), Washington, DC, USA, October 2011, pp. 791–796.
-
-
11)
-
9. Black, M.J., Jepson, A.D.: ‘Eigentracking: robust matching and tracking of articulated objects using a view-based representation’, Int. J. Comput. Vis., 1998, 26, (1), pp. 63–84.
-
-
12)
-
13. Avidan, S.: ‘Ensemble tracking’, IEEE Trans. Pattern Anal. Mach. Intell., 2007, 29, (2), pp. 261–271.
-
-
13)
-
32. Song, S., Xiao, J.: ‘Tracking revisited using rgbd camera: unified benchmark and baselines’. The IEEE Int. Conf. Computer Vision, Sydney, Australia, December 2013, pp. 233–240.
-
-
14)
-
38. Hare, S., Saffari, A., Torr, P.H.: ‘Struck: structured output tracking with kernels’, IEEE Trans. Pattern Anal. Mach. Intell., 2016, 38, (10), pp. 263–270.
-
-
15)
-
40. Dalal, N., Triggs, B.: ‘Histograms of oriented gradients for human detection’. The IEEE Conf. Computer Vision and Pattern Recognition, San Diego, USA, June 2005, pp. 886–893.
-
-
16)
-
2. Wu, Y., Lim, J., Yang, M.H.: ‘Online object tracking: a benchmark’. The IEEE Conf. Computer Vision and Pattern Recognition, Portland, USA, June 2013, pp. 2411–2418.
-
-
17)
-
12. Ross, D.A., Lim, J., Lin, R.S., et al: ‘Incremental learning for robust visual tracking’, Int. J. Comput. Vis., 2008, 77, (1–3), pp. 125–141.
-
-
18)
-
7. Hong, Z., Mei, X., Prokhorov, D., et al: ‘Tracking via robust multi-task multi-view joint sparse representation’. The IEEE Int. Conf. Computer Vision, Sydney, Australia, December 2013, pp. 649–656.
-
-
19)
-
24. Yang, R., Zhang, Z.: ‘Model-based head pose tracking with stereovision’. Fifth IEEE Int. Conf. Automatic Face and Gesture Recognition, Washington, DC, USA, May 2002, pp. 255–260.
-
-
20)
-
22. Zhang, T., Liu, S., Xu, C., et al: ‘Structural sparse tracking’. The IEEE Conf. Computer Vision and Pattern Recognition, Boston, USA, June 2015, pp. 150–158.
-
-
21)
-
4. Mei, X., Ling, H.: ‘Robust visual tracking using l1 minimization’. The IEEE Int. Conf. Computer Vision, Kyoto, Japan, September 2009, pp. 1436–1443.
-
-
22)
-
35. Chen, X., Pan, W.K., Kwok, J.T., et al: ‘Accelerated gradient method for multi-task sparse learning problem’. The IEEE Int. Conf. Data Mining, Miami, USA, December 2009, pp. 746–751.
-
-
23)
-
20. Zhang, T., Liu, S., Ahuja, N., et al: ‘Robust visual tracking via consistent low-rank sparse learning’, Int. J. Comput. Vis., 2015, 111, (2), pp. 171–190.
-
-
24)
-
23. Nirei, K., Saito, H., Mochimaru, M., et al: ‘Human hand tracking from binocular image sequences’. The IEEE IECON 22nd Int. Conf. Industrial Electronics, Control, and Instrumentation, Taipei, Taiwan, August 1996, pp. 297–302.
-
-
25)
-
18. Bao, C., Wu, Y., Ling, H., et al: ‘Real time robust l1 tracker using accelerated proximal gradient approach’. The IEEE Conf. Computer Vision and Pattern Recognition, Providence, RI, USA, June 2012, pp. 1830–1837.
-
-
26)
-
33. Zhang, Z.: ‘Iterative point matching for registration of free-form curves and surfaces’, Int. J. Comput. Vis., 1994, 13, (2), pp. 119–152.
-
-
27)
-
27. Rabe, C., Franke, U., Gehrig, S.: ‘Fast detection of moving objects in complex scenarios’. Intelligent Vehicles Symp., Istanbul, Turkey, June 2007, pp. 398–403.
-
-
28)
-
39. Henriques, J.F., Caseiro, R., Martins, P., et al: ‘High-speed tracking with kernelized correlation filters’, IEEE Trans. Pattern Anal. Mach. Intell., 2015, 37, (3), pp. 583–596.
-
-
29)
-
11. Porikli, F., Tuzel, O., Meer, P.: ‘Covariance tracking using model update based on lie algebra’. The IEEE Conf. Computer Vision and Pattern Recognition, New York, USA, July 2006, pp. 728–735.
-
-
30)
-
31. Sivaraman, S., Trivedi, M.M.: ‘Looking at vehicles on the road: A survey of vision-based vehicle detection, tracking, and behavior analysis’, IEEE Trans. Intell. Transp. Syst., 2013, 14, (4), pp. 1773–1795.
-
-
31)
-
16. Yin, Z., Collins, R.T.: ‘Object tracking and detection after occlusion via numerical hybrid local and global mode-seeking’. The IEEE Conf. Computer Vision and Pattern Recognition, Anchorage, USA, June 2008, pp. 1–8.
-
-
32)
-
6. Zhang, T., Ghanem, B., Liu, S., et al: ‘Robust visual tracking via multi-task sparse learning’. The IEEE Conf. Computer Vision and Pattern Recognition, Providence, RI, UAS, June 2012, pp. 2042–2049.
-
-
33)
-
17. Mei, X., Ling, H., Wu, Y., et al: ‘Minimum error bounded efficient l1 tracker with occlusion detection’. The IEEE Conf. Computer Vision and Pattern Recognition, Providence, RI, USA, June 2011, pp. 1257–1264.
-
-
34)
-
34. Wan, L.C., Sebastian, P., Voon, Y.V.: ‘Stereo vision tracking system’. Int. Conf. Future Computer and Communication, Kuala Lumpar, Malaysia, April 2009, pp. 487–491.
-
-
35)
-
37. Peng, Y., Ganesh, A., Wright, J., et al: ‘Rasl: Robust alignment by sparse and low-rank decomposition for linearly correlated images’, IEEE Trans. Pattern Anal. Mach. Intell., 2012, 34, (11), pp. 2233–2246.
-
-
36)
-
5. Mei, X., Ling, H.: ‘Robust visual tracking and vehicle classification via sparse representation’, IEEE Trans. Pattern Anal. Mach. Intell., 2011, 33, (11), pp. 2259–2272.
-
-
37)
-
41. Ojala, T., Pietikäinen, M., Mäenpää, T.: ‘Gray scale and rotation invariant texture classification with local binary patterns’. European Conf. Computer Vision, Dublin, Ireland, June 2000, pp. 404–420.
-
-
38)
-
36. Boyd, S., Parikh, N., Chu, E., et al: ‘Distributed optimization and statistical learning via the alternating direction method of multipliers’, Found. Trends R in Mach. Learn., 2011, 3, (1), pp. 1–122.
-
-
39)
-
8. Lan, X., Ma, A., Yuen, P.: ‘Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation’. The IEEE Conf. Computer Vision and Pattern Recognition, Columbus, Ohio, USA, June 2014, pp. 1194–1201.
-
-
40)
-
25. Bae, K.H., Koo, J.S., Kim, E.S.: ‘A new stereo object tracking system using disparity motion vector’, Opt. Commun., 2003, 221, (1), pp. 23–35.
-
-
41)
-
15. Williams, O., Blake, A., Cipolla, R.: ‘Sparse Bayesian learning for efficient visual tracking’, IEEE Trans. Pattern Anal. Mach. Intell., 2005, 27, (8), pp. 1292–1304.
-
-
1)