scholarly journals Cooperative Multi-Agent Reinforcement Learning with Conversation Knowledge for Dialogue Management

2020 ◽  
Vol 10 (8) ◽  
pp. 2740
Author(s):  
Shuyu Lei ◽  
Xiaojie Wang ◽  
Caixia Yuan

Dialogue management plays a vital role in task-oriented dialogue systems, which has become an active area of research in recent years. Despite the promising results brought from deep reinforcement learning, most of the studies need to develop a manual user simulator additionally. To address the time-consuming development of simulator policy, we propose a multi-agent dialogue model where an end-to-end dialogue manager and a user simulator are optimized simultaneously. Different from prior work, we optimize the two-agents from scratch and apply the reward shaping technology based on adjacency pairs constraints in conversational analysis to speed up learning and to avoid the derivation from normal human-human conversation. In addition, we generalize the one-to-one learning strategy to one-to-many learning strategy, where a dialogue manager can be concurrently optimized with various user simulators, to improve the performance of trained dialogue manager. The experimental results show that one-to-one agents trained with adjacency pairs constraints can converge faster and avoid derivation. In cross-model evaluation with human users involved, the dialogue manager trained in one-to-many strategy achieves the best performance.

Author(s):  
Omar Sami Oubbati ◽  
Mohammed Atiquzzaman ◽  
Abderrahmane Lakas ◽  
Abdullah Baz ◽  
Hosam Alhakami ◽  
...  

2017 ◽  
Vol 26 (01) ◽  
pp. 1760009 ◽  
Author(s):  
Guillaume Dubuisson Duplessis ◽  
Alexandre Pauchet ◽  
Nathalie Chaignaud ◽  
Jean-Philippe Kotowicz

Our work aims at designing a dialogue manager dedicated to agents that interact with humans. In this article, we investigate how dialogue patterns at the dialogue act level extracted from Human-Human interactions can be fruitfully used by a software agent to interact with a human.We show how these patterns can be leveraged via a dialogue game structure in order to benefit to the dialogue management process of an agent. We describe how empirically specified dialogue games can be employed on both interpretative and generative levels of dialogue management. We present Dogma, an open-source module that can be used by an agent to manage its conventional communicative behaviour. We show that our library of dialogue games can be used into Dogma to generate fragments of dialogue that are strongly coherent from a human perspective.


Measurement ◽  
2021 ◽  
pp. 109955
Author(s):  
Lei Xi ◽  
Mengmeng Sun ◽  
Huan Zhou ◽  
Yanchun Xu ◽  
Junnan Wu ◽  
...  

2019 ◽  
Author(s):  
Alexandros Papangelis ◽  
Yi-Chia Wang ◽  
Piero Molino ◽  
Gokhan Tur

Author(s):  
Tulika Saha ◽  
Dhawal Gupta ◽  
Sriparna Saha ◽  
Pushpak Bhattacharyya

Building Virtual Agents capable of carrying out complex queries of the user involving multiple intents of a domain is quite a challenge, because it demands that the agent manages several subtasks simultaneously. This article presents a universal Deep Reinforcement Learning framework that can synthesize dialogue managers capable of working in a task-oriented dialogue system encompassing various intents pertaining to a domain. The conversation between agent and user is broken down into hierarchies, to segregate subtasks pertinent to different intents. The concept of Hierarchical Reinforcement Learning, particularly options , is used to learn policies in different hierarchies that operates in distinct time steps to fulfill the user query successfully. The dialogue manager comprises top-level intent meta-policy to select among subtasks or options and a low-level controller policy to pick primitive actions to communicate with the user to complete the subtask provided to it by the top-level policy in varying intents of a domain. The proposed dialogue management module has been trained in a way such that it can be reused for any language for which it has been developed with little to no supervision. The developed system has been demonstrated for “Air Travel” and “Restaurant” domain in English and Hindi languages. Empirical results determine the robustness and efficacy of the learned dialogue policy as it outperforms several baselines and a state-of-the-art system.


Author(s):  
Yue Hu ◽  
Juntao Li ◽  
Xi Li ◽  
Gang Pan ◽  
Mingliang Xu

As an important and challenging problem in artificial intelligence (AI) game playing, StarCraft micromanagement involves a dynamically adversarial game playing process with complex multi-agent control within a large action space. In this paper, we propose a novel knowledge-guided agent-tactic-aware learning scheme, that is, opponent-guided tactic learning (OGTL), to cope with this micromanagement problem. In principle, the proposed scheme takes a two-stage cascaded learning strategy which is capable of not only transferring the human tactic knowledge from the human-made opponent agents to our AI agents but also improving the adversarial ability. With the power of reinforcement learning, such a knowledge-guided agent-tactic-aware scheme has the ability to guide the AI agents to achieve high winning-rate performances while accelerating the policy exploration process in a tactic-interpretable fashion. Experimental results demonstrate the effectiveness of the proposed scheme against the state-of-the-art approaches in several benchmark combat scenarios.


Author(s):  
Hao Jiang ◽  
Dianxi Shi ◽  
Chao Xue ◽  
Yajie Wang ◽  
Gongju Wang ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document