http://iet.metastore.ingenta.com
1887

Joint prominent expression feature regions in auxiliary task learning network for facial expression recognition

Joint prominent expression feature regions in auxiliary task learning network for facial expression recognition

For access to this article, please select a purchase option:

Buy article PDF
£12.50
(plus tax if applicable)
Buy Knowledge Pack
10 articles for £75.00
(plus taxes if applicable)

IET members benefit from discounts to all IET publications and free access to E&T Magazine. If you are an IET member, log in to your account and the discounts will automatically be applied.

Learn more about IET membership 

Recommend Title Publication to library

You must fill out fields marked with: *

Librarian details
Name:*
Email:*
Your details
Name:*
Email:*
Department:*
Why are you recommending this title?
Select reason:
 
 
 
 
 
Electronics Letters — Recommend this title to your library

Thank you

Your recommendation has been sent to your librarian.

The key issue for facial expression recognition (FER) is to concentrate on the prominent expression feature regions where the expression changes. In this Letter, the authors propose a novel and effective FER framework jointing prominent expression feature regions in an auxiliary task learning network (ATLN). The proposed approach consists of two deep learning neural networks, one of which is the main network whose inputs are complete face images, and the other is the auxiliary learning network whose inputs are pre-processed face images containing prominent expression feature regions. The main network structure of the ATLN improves its ability to focus on regions with prominent expressions changing by sharing parameters with their auxiliary network structure. They carry out experiments on two public facial expression databases, namely, CK+ and MMI. Experimental results demonstrate the superior performance of proposed method.

References

    1. 1)
    2. 2)
    3. 3)
      • 3. Zheng, W., Tang, H., Lin, Z., et al: ‘Emotion recognition from arbitrary view facial images’. European Conf. on Computer Vision (ECCV 2010), Heraklion, Crete, Greece, 2010, pp. 490503.
    4. 4)
      • 4. Yolcu, G., Oztel, I., Kazan, S., et al: ‘Deep learning-based facial expression recognition for monitoring neurological disorders’. IEEE Int. Conf. on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, November 2017, pp. 16521657.
    5. 5)
      • 5. Mollahosseini, A., Chan, D., Mahoor, M.H.: ‘Going deeper in facial expression recognition using deep neural networks’. IEEE Winter Conf. on Applications of Computer Vision (WACV), Lake Placid, NY, USA, March 2016, pp. 110.
    6. 6)
      • 6. Liu, M., Shan, S., Wang, R., et al: ‘Learning expressionlets on spatio-temporal manifold for dynamic facial expression recognition’. IEEE Conf. on Computer Vision and Pattern Recognition, Columbus, OH, USA, June 2014, pp. 17491756.
    7. 7)
      • 7. Zhang, C., Wang, P., Chen, K., et al: ‘Identity-aware convolutional neural networks for facial expression recognition’, J. Syst. Eng. Electron., 2017, 28, (4), pp. 784792.
    8. 8)
      • 8. Cai, J., Meng, Z., Khan, A., et al: ‘Island loss for learning discriminative features in facial expression recognition’, 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition, Xi'an, China, May 2018, pp. 302309.
    9. 9)
      • 9. Li, S., Deng, D., Du, J.: ‘Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild’. IEEE Conf. on Computer Vision and Pattern Recognition, Honolulu, HI, USA, July 2017, pp. 25842593.
    10. 10)
      • 10. Jung, H., Lee, S., Yim, J., et al: ‘Joint fine-tuning in deep neural networks for facial expression recognition’, 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, December 2015, pp. 29832991.
    11. 11)
      • 11. Sikka, K., Sharma, G., Bartlett, M.: ‘LOMo: latent ordinal model for facial analysis in videos’. IEEE Conf. on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, June 2016, pp. 55805589.
    12. 12)
http://iet.metastore.ingenta.com/content/journals/10.1049/el.2018.7235
Loading

Related content

content/journals/10.1049/el.2018.7235
pub_keyword,iet_inspecKeyword,pub_concept
6
6
Loading
This is a required field
Please enter a valid email address