papers AI Learner
The Github is limit! Click to go to the new site.

Convolutional Attention Networks for Multimodal Emotion Recognition from Speech and Text Data

2019-03-08
Chan Woo Lee, Kyu Ye Song, Jihoon Jeong, Woo Yong Choi

Abstract

Emotion recognition has become a popular topic of interest, especially in the field of human computer interaction. Previous works involve unimodal analysis of emotion, while recent efforts focus on multi-modal emotion recognition from vision and speech. In this paper, we propose a new method of learning about the hidden representations between just speech and text data using convolutional attention networks. Compared to the shallow model which employs simple concatenation of feature vectors, the proposed attention model performs much better in classifying emotion from speech and text data contained in the CMU-MOSEI dataset.

Abstract (translated by Google)
URL

http://arxiv.org/abs/1805.06606

PDF

http://arxiv.org/e-print/1805.06606


Similar Posts

Comments