Planning and Coordination in Multiagent Environments

Author(s):  
Alejandro Torreño ◽  
Eva Onaindia ◽  
Vicente Botti
Author(s):  
Takuji Watanabe ◽  
◽  
Kazuteru Miyazaki ◽  
Hiroaki Kobayashi ◽  
◽  
...  

The penalty avoiding rational policy making algorithm (PARP) [1] previously improved to save memory and cope with uncertainty, i.e., IPARP [2], requires that states be discretized in real environments with continuous state spaces, using function approximation or some other method. Especially, in PARP, a method that discretizes state using a basis functions is known [3]. Because this creates a new basis function based on the current input and its next observation, however, an unsuitable basis function may be generated in some asynchronous multiagent environments. We therefore propose a uniform basis function and range extent of the basis function is estimated before learning. We show the effectiveness of our proposal using a soccer game task called “Keepaway.”


2002 ◽  
Vol 14 (2) ◽  
pp. 281-295 ◽  
Author(s):  
A. Nur Zincir-Heywood ◽  
M.I. Heywood ◽  
C.R. Chatwin

Sign in / Sign up

Export Citation Format

Share Document