Trimodal Attention Module for Multimodal Sentiment Analysis (Student Abstract)
2020 ◽
Vol 34
(10)
◽
pp. 13803-13804
Keyword(s):
In our research, we propose a new multimodal fusion architecture for the task of sentiment analysis. The 3 modalities used in this paper are text, audio and video. Most of the current methods deal with either a feature level or a decision level fusion. In contrast, we propose an attention-based deep neural network and a training approach to facilitate both feature and decision level fusion. Our network effectively leverages information across all three modalities using a 2 stage fusion process. We test our network on the individual utterance based contextual information extracted from the CMU-MOSI Dataset. A comparison is drawn between the state-of-the-art and our network.
Keyword(s):
2011 ◽
Vol 9
(3)
◽
pp. 031002-31005
◽
2019 ◽
Vol 33
◽
pp. 6851-6858
◽
Keyword(s):
2020 ◽
Vol 8
(5)
◽
pp. 2522-2527
Keyword(s):
2004 ◽
pp. 307-317
◽
Keyword(s):