A Non-Task-Oriented Dialogue System Controlling the Utterance Length

Author(s):  
Kazuki Isoshima ◽  
Masafumi Hagiwara
Author(s):  
Yan Peng ◽  
Penghe Chen ◽  
Yu Lu ◽  
Qinggang Meng ◽  
Qi Xu ◽  
...  

2019 ◽  
Vol 23 (3) ◽  
pp. 1989-2002 ◽  
Author(s):  
Haotian Xu ◽  
Haiyun Peng ◽  
Haoran Xie ◽  
Erik Cambria ◽  
Liuyang Zhou ◽  
...  

2018 ◽  
Author(s):  
Yuya Chiba ◽  
Takashi Nose ◽  
Taketo Kase ◽  
Mai Yamanaka ◽  
Akinori Ito

2021 ◽  
Vol 12 (2) ◽  
pp. 1-33
Author(s):  
Mauajama Firdaus ◽  
Nidhi Thakur ◽  
Asif Ekbal

Multimodality in dialogue systems has opened up new frontiers for the creation of robust conversational agents. Any multimodal system aims at bridging the gap between language and vision by leveraging diverse and often complementary information from image, audio, and video, as well as text. For every task-oriented dialog system, different aspects of the product or service are crucial for satisfying the user’s demands. Based upon the aspect, the user decides upon selecting the product or service. The ability to generate responses with the specified aspects in a goal-oriented dialogue setup facilitates user satisfaction by fulfilling the user’s goals. Therefore, in our current work, we propose the task of aspect controlled response generation in a multimodal task-oriented dialog system. We employ a multimodal hierarchical memory network for generating responses that utilize information from both text and images. As there was no readily available data for building such multimodal systems, we create a Multi-Domain Multi-Modal Dialog (MDMMD++) dataset. The dataset comprises the conversations having both text and images belonging to the four different domains, such as hotels, restaurants, electronics, and furniture. Quantitative and qualitative analysis on the newly created MDMMD++ dataset shows that the proposed methodology outperforms the baseline models for the proposed task of aspect controlled response generation.


Sign in / Sign up

Export Citation Format

Share Document