Robust video tracking algorithm: a multi-feature fusion approach

Robust video tracking algorithm: a multi-feature fusion approach

For access to this article, please select a purchase option:

Buy eFirst article PDF
(plus tax if applicable)
Buy Knowledge Pack
10 articles for £75.00
(plus taxes if applicable)

IET members benefit from discounts to all IET publications and free access to E&T Magazine. If you are an IET member, log in to your account and the discounts will automatically be applied.

Learn more about IET membership 

Recommend to library

You must fill out fields marked with: *

Librarian details
Your details
Why are you recommending this title?
Select reason:
— Recommend this title to your library

Thank you

Your recommendation has been sent to your librarian.

This study proposes a novel robust video tracking algorithm consists of target detection, multi-feature fusion, and extended Camshift. Firstly, a novel target detection method that integrates Canny edge operator, three-frame difference, and improved Gaussian mixture model (IGMM)-based background modelling is provided to detect targets. The IGMM-based background modelling divides video frames into meshes to avoid pixel-wise processing. In addition, the output of the target detection is utilised to initialise the IGMM and to accelerate the convergence of iterations. Secondly, low-dimensional regional covariance matrices are introduced to describe video targets by fusing multiple features like pixel location, colour index, rotation and scale invariant features as well as uniform local binary patterns, and directional derivatives. Thirdly, an extended Camshift based on adaptive kernel bandwidth and robust H state estimation is proposed to predict the states of fast moving targets and to reduce the mean shift iterations. Finally, the effectiveness of the proposed tracking algorithm is demonstrated via experiments.

Related content

This is a required field
Please enter a valid email address