Multi-Agent Motion Control in Cluttered and Noisy Environments

2013 ◽  
Vol 8 (1) ◽  
pp. 32-46 ◽  
Author(s):  
Hung Manh La ◽  
Weihua Sheng
PAMM ◽  
2012 ◽  
Vol 12 (1) ◽  
pp. 733-734 ◽  
Author(s):  
Axel Hackbarth ◽  
Edwin Kreuzer ◽  
Andrew Gray

Entropy ◽  
2021 ◽  
Vol 23 (9) ◽  
pp. 1133
Author(s):  
Shanzhi Gu ◽  
Mingyang Geng ◽  
Long Lan

The aim of multi-agent reinforcement learning systems is to provide interacting agents with the ability to collaboratively learn and adapt to the behavior of other agents. Typically, an agent receives its private observations providing a partial view of the true state of the environment. However, in realistic settings, the harsh environment might cause one or more agents to show arbitrarily faulty or malicious behavior, which may suffice to allow the current coordination mechanisms fail. In this paper, we study a practical scenario of multi-agent reinforcement learning systems considering the security issues in the presence of agents with arbitrarily faulty or malicious behavior. The previous state-of-the-art work that coped with extremely noisy environments was designed on the basis that the noise intensity in the environment was known in advance. However, when the noise intensity changes, the existing method has to adjust the configuration of the model to learn in new environments, which limits the practical applications. To overcome these difficulties, we present an Attention-based Fault-Tolerant (FT-Attn) model, which can select not only correct, but also relevant information for each agent at every time step in noisy environments. The multihead attention mechanism enables the agents to learn effective communication policies through experience concurrent with the action policies. Empirical results showed that FT-Attn beats previous state-of-the-art methods in some extremely noisy environments in both cooperative and competitive scenarios, much closer to the upper-bound performance. Furthermore, FT-Attn maintains a more general fault tolerance ability and does not rely on the prior knowledge about the noise intensity of the environment.


Robotica ◽  
2001 ◽  
Vol 19 (6) ◽  
pp. 611-617 ◽  
Author(s):  
Amar Khoukhi

In this paper fhe problem of motion control of a biped is considered. We develop a new method based on multi-agent associated Neural AIGLS (On-line Augmented Integration of Gradient and Last Sguare method) – RSPN (Recursive Stochastic Petri Nets) strategy. This method deals with organization and coordination aspects in an intelligent modeling of human motion. We propose a cooperative multi-agent model. Based on this model, we develop a control kernel named IMCOK (Intelligent Motion COntrol Kernel) which consists of a controller, a coordinator and an executor of different cycles of the motion of the biped. When walking, IMCOK receives messages and sends offers. A Decision Making of Actions (DMA) is developed at the supervisor level. The articulator agents partially planify the motion of the associated non-articulator agents. The system is hybrid and distributed functionally. The learning of the biped is performed using an On-line Augmented Integration of Gradient and Last Sguare Neural Networks based algorithm. In the conflictual situations of sending or receiving messages by the managers of MABS we apply a new strategy: Recursive Stochastic Petri Nets (RSPN). This module is fundamental in the On-line information processing between agents. It allows particularly the Recursive strategy concept. Cognitive agents communicate with reactive (non-articulator) agents in order to generate the motion.


Sign in / Sign up

Export Citation Format

Share Document