Multimodal learning for facial expression recognition

Pattern Recognition(2015)

引用 112|浏览110
暂无评分
摘要
In this paper, multimodal learning for facial expression recognition (FER) is proposed. The multimodal learning method makes the first attempt to learn the joint representation by considering the texture and landmark modality of facial images, which are complementary with each other. In order to learn the representation of each modality and the correlation and interaction between different modalities, the structured regularization (SR) is employed to enforce and learn the modality-specific sparsity and density of each modality, respectively. By introducing SR, the comprehensiveness of the facial expression is fully taken into consideration, which can not only handle the subtle expression but also perform robustly to different input of facial images. With the proposed multimodal learning network, the joint representation learning from multimodal inputs will be more suitable for FER. Experimental results on the CK+ and NVIE databases demonstrate the superiority of our proposed method. HighlightsMultimodal learning for facial expression recognition (FER) is proposed.The first attempt to do FER from the joint representation of texture and landmarks.The multimodal structure combines feature extraction and classification together.Structured regularization is used to enforce the sparsity of different modalities.
更多
查看译文
关键词
Multimodal learning,Facial expression recognition,Texture,Landmark
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要