scholarly journals Multiagent Cooperative Learning Strategies for Pursuit-Evasion Games

2015 ◽  
Vol 2015 ◽  
pp. 1-13 ◽  
Author(s):  
Jong Yih Kuo ◽  
Hsiang-Fu Yu ◽  
Kevin Fong-Rey Liu ◽  
Fang-Wen Lee

This study examines the pursuit-evasion problem for coordinating multiple robotic pursuers to locate and track a nonadversarial mobile evader in a dynamic environment. Two kinds of pursuit strategies are proposed, one for agents that cooperate with each other and the other for agents that operate independently. This work further employs the probabilistic theory to analyze the uncertain state information about the pursuers and the evaders and uses case-based reasoning to equip agents with memories and learning abilities. According to the concepts of assimilation and accommodation, both positive-angle and bevel-angle strategies are developed to assist agents in adapting to their environment effectively. The case study analysis uses the Recursive Porous Agent Simulation Toolkit (REPAST) to implement a multiagent system and demonstrates superior performance of the proposed approaches to the pursuit-evasion game.

Entropy ◽  
2021 ◽  
Vol 23 (11) ◽  
pp. 1433
Author(s):  
Kaifang Wan ◽  
Dingwei Wu ◽  
Yiwei Zhai ◽  
Bo Li ◽  
Xiaoguang Gao ◽  
...  

A pursuit–evasion game is a classical maneuver confrontation problem in the multi-agent systems (MASs) domain. An online decision technique based on deep reinforcement learning (DRL) was developed in this paper to address the problem of environment sensing and decision-making in pursuit–evasion games. A control-oriented framework developed from the DRL-based multi-agent deep deterministic policy gradient (MADDPG) algorithm was built to implement multi-agent cooperative decision-making to overcome the limitation of the tedious state variables required for the traditionally complicated modeling process. To address the effects of errors between a model and a real scenario, this paper introduces adversarial disturbances. It also proposes a novel adversarial attack trick and adversarial learning MADDPG (A2-MADDPG) algorithm. By introducing an adversarial attack trick for the agents themselves, uncertainties of the real world are modeled, thereby optimizing robust training. During the training process, adversarial learning was incorporated into our algorithm to preprocess the actions of multiple agents, which enabled them to properly respond to uncertain dynamic changes in MASs. Experimental results verified that the proposed approach provides superior performance and effectiveness for pursuers and evaders, and both can learn the corresponding confrontational strategy during training.


2020 ◽  
Vol 53 (2) ◽  
pp. 14882-14887
Author(s):  
Yuan Chai ◽  
Jianjun Luo ◽  
Mingming Wang ◽  
Min Yu

Complexity ◽  
2020 ◽  
Vol 2020 ◽  
pp. 1-13
Author(s):  
Xiao Liang ◽  
Honglun Wang ◽  
Haitao Luo

The UAV/UGV heterogeneous system combines the air superiority of UAV (unmanned aerial vehicle) and the ground superiority of UGV (unmanned ground vehicle). The system can complete a series of complex tasks and one of them is pursuit-evasion decision, so a collaborative strategy of UAV/UGV heterogeneous system is proposed to derive a pursuit-evasion game in complex three-dimensional (3D) polygonal environment, which is large enough but with boundary. Firstly, the system and task hypothesis are introduced. Then, an improved boundary value problem (BVP) is used to unify the terrain data of decision and path planning. Under the condition that the evader knows the position of collaborative pursuers at any time but pursuers just have a line-of-sight view, a worst case is analyzed and the strategy between the evader and pursuers is studied. According to the state of evader, the strategy of collaborative pursuers is discussed in three situations: evader is in the visual field of pursuers, evader just disappears from the visual field of pursuers, and the position of evader is completely unknown to pursuers. The simulation results show that the strategy does not guarantee that the pursuers will win the game in complex 3D polygonal environment, but it is optimal in the worst case.


1999 ◽  
Vol 09 (04n05) ◽  
pp. 471-493 ◽  
Author(s):  
LEONIDAS J. GUIBAS ◽  
JEAN-CLAUDE LATOMBE ◽  
STEVEN M. LAVALLE ◽  
DAVID LIN ◽  
RAJEEV MOTWANI

This paper addresses the problem of planning the motion of one or more pursuers in a polygonal environment to eventually "see" an evader that is unpredictable, has unknown initial position, and is capable of moving arbitrarily fast. This problem was first introduced by Suzuki and Yamashita. Our study of this problem is motivated in part by robotics applications, such as surveillance with a mobile robot equipped with a camera that must find a moving target in a cluttered workspace. A few bounds are introduced, and a complete algorithm is presented for computing a successful motion strategy for a single pursuer. For simply-connected free spaces, it is shown that the minimum number of pursuers required is Θ( lg  n). For multiply-connected free spaces, the bound is [Formula: see text] pursuers for a polygon that has n edges and h holes. A set of problems that are solvable by a single pursuer and require a linear number of recontaminations is shown. The complete algorithm searches a finite graph that is constructed on the basis of critical information changes. It has been implemented and computed examples are shown.


Sign in / Sign up

Export Citation Format

Share Document