Title
Multi-clue fusion for emotion recognition in the wild.
Abstract
In the past three years, Emotion Recognition in the Wild (EmotiW) Grand Challenge has drawn more and more attention due to its huge potential applications. In the fourth challenge, aimed at the task of video based emotion recognition, we propose a multi-clue emotion fusion (MCEF) framework by modeling human emotion from three mutually complementary sources, facial appearance texture, facial action, and audio. To extract high-level emotion features from sequential face images, we employ a CNN-RNN architecture, where face image from each frame is first fed into the fine-tuned VGG-Face network to extract face feature, and then the features of all frames are sequentially traversed in a bidirectional RNN so as to capture dynamic changes of facial textures. To attain more accurate facial actions, a facial landmark trajectory model is proposed to explicitly learn emotion variations of facial components. Further, audio signals are also modeled in a CNN framework by extracting low-level energy features from segmented audio clips and then stacking them as an image-like map. Finally, we fuse the results generated from three clues to boost the performance of emotion recognition. Our proposed MCEF achieves an overall accuracy of 56.66% with a large improvement of 16.19% with respect to the baseline.
Year
DOI
Venue
2018
10.1016/j.neucom.2018.03.068
Neurocomputing
Keywords
DocType
Volume
Emotion recognition,Convolutional neural network (CNN),Facial landmark action,Multi-cue fusion
Journal
309
ISSN
Citations 
PageRank 
0925-2312
15
0.63
References 
Authors
21
7
Name
Order
Citations
PageRank
Jingwei Yan1683.44
Wenming Zheng2124080.70
Zhen Cui358041.43
Chuangao Tang4284.25
Tong Zhang5494.91
Yuan Zong6150.63
Ning Sun711813.20