A Survey on Automatic Multimodal Emotion Recognition in the Wild

Author(s):  
Garima Sharma ◽  
Abhinav Dhall
Author(s):  
Jing Chen ◽  
Chenhui Wang ◽  
Kejun Wang ◽  
Chaoqun Yin ◽  
Cong Zhao ◽  
...  

2021 ◽  
Vol 25 (4) ◽  
pp. 1031-1045
Author(s):  
Helang Lai ◽  
Keke Wu ◽  
Lingli Li

Emotion recognition in conversations is crucial as there is an urgent need to improve the overall experience of human-computer interactions. A promising improvement in this field is to develop a model that can effectively extract adequate contexts of a test utterance. We introduce a novel model, termed hierarchical memory networks (HMN), to address the issues of recognizing utterance level emotions. HMN divides the contexts into different aspects and employs different step lengths to represent the weights of these aspects. To model the self dependencies, HMN takes independent local memory networks to model these aspects. Further, to capture the interpersonal dependencies, HMN employs global memory networks to integrate the local outputs into global storages. Such storages can generate contextual summaries and help to find the emotional dependent utterance that is most relevant to the test utterance. With an attention-based multi-hops scheme, these storages are then merged with the test utterance using an addition operation in the iterations. Experiments on the IEMOCAP dataset show our model outperforms the compared methods with accuracy improvement.


2021 ◽  
pp. 1-1
Author(s):  
Shao-Yen Tseng ◽  
Shrikanth Narayanan ◽  
Panayiotis Georgiou

2021 ◽  
Vol 19 (2) ◽  
Author(s):  
Dong Liu ◽  
Longxi Chen ◽  
Zhiyong Wang ◽  
Guangqiang Diao

2021 ◽  
Author(s):  
Puneet Kumar ◽  
Vedanti Khokher ◽  
Yukti Gupta ◽  
Balasubramanian Raman

2021 ◽  
Author(s):  
Yibo Huang ◽  
Hongqian Wen ◽  
Linbo Qing ◽  
Rulong Jin ◽  
Leiming Xiao

2018 ◽  
Vol E101.D (8) ◽  
pp. 2092-2100
Author(s):  
Nurul LUBIS ◽  
Dessi LESTARI ◽  
Sakriani SAKTI ◽  
Ayu PURWARIANTI ◽  
Satoshi NAKAMURA

Sign in / Sign up

Export Citation Format

Share Document