An elastic net-regularized HMAX model of visual processing
An elastic net-regularized HMAX model of visual processing
- Author(s): A. Alameer ; G. Ghazaeil ; P. Degenaar ; K. Nazarpour
- DOI: 10.1049/cp.2015.1753
For access to this article, please select a purchase option:
Buy conference paper PDF
Buy Knowledge Pack
IET members benefit from discounts to all IET publications and free access to E&T Magazine. If you are an IET member, log in to your account and the discounts will automatically be applied.
2nd IET International Conference on Intelligent Signal Processing 2015 (ISP) — Recommend this title to your library
Thank you
Your recommendation has been sent to your librarian.
- Author(s): A. Alameer ; G. Ghazaeil ; P. Degenaar ; K. Nazarpour Source: 2nd IET International Conference on Intelligent Signal Processing 2015 (ISP), 2015 page ()
- Conference: 2nd IET International Conference on Intelligent Signal Processing 2015 (ISP)
- « Previous article
- Table of contents
- Next article »
- DOI: 10.1049/cp.2015.1753
- ISBN: 978-1-78561-136-0
- Location: London, UK
- Conference date: 1-2 Dec. 2015
- Format: PDF
The hierarchical MAX (HMAX) model of human visual system has been used in robotics and autonomous systems widely. However, there is still a stark gap between human and robotic vision in observing the environment and intelligently categorizing the objects. Therefore, improving models such as the HMAX is still topical. In this work, in order to enhance the performance of HMAX in an object recognition task, we augmented it using an elastic net-regularised dictionary learning approach. We used the notion of sparse coding in the S layers of the HMAX model to extract mid- and high-level, i.e. abstract, features from input images. In addition, we used spatial pyramid pooling (SPP) at the output of higher layers to create a fixed feature vectors before feeding them into a softmax classifier. In our model, the sparse coefficients calculated by the elastic net-regularised dictionary learning algorithm were used to train and test the model. With this setup, we achieved a classification accuracy of 82.6387%∓3.7183% averaged across 5-folds which is significantly better than that achieved with the original HMAX.
Inspec keywords: feature extraction; visual perception; image classification; robot vision; object recognition
Subjects: Computer vision and image processing techniques; Robotics; Image recognition
Related content
content/conferences/10.1049/cp.2015.1753
pub_keyword,iet_inspecKeyword,pub_concept
6
6
