markov games
Recently Published Documents


TOTAL DOCUMENTS

159
(FIVE YEARS 39)

H-INDEX

20
(FIVE YEARS 2)

2021 ◽  
Vol 32 (1) ◽  
pp. 13
Author(s):  
Shirin Kordnoori ◽  
Hamidreza Mostafaei ◽  
Mohammadmohsen Ostadrahimi ◽  
Saeed Agha Banihashemi

2021 ◽  
Vol 71 ◽  
pp. 925-951
Author(s):  
Justin Fu ◽  
Andrea Tacchetti ◽  
Julien Perolat ◽  
Yoram Bachrach

A core question in multi-agent systems is understanding the motivations for an agent's actions based on their behavior. Inverse reinforcement learning provides a framework for extracting utility functions from observed agent behavior, casting the problem as finding domain parameters which induce such a behavior from rational decision makers.  We show how to efficiently and scalably extend inverse reinforcement learning to multi-agent settings, by reducing the multi-agent problem to N single-agent problems while still satisfying rationality conditions such as strong rationality. However, we observe that rewards learned naively tend to lack insightful structure, which causes them to produce undesirable behavior when optimized in games with different players from those encountered during training. We further investigate conditions under which rewards or utility functions can be precisely identified, on problem domains such as normal-form and Markov games, as well as auctions, where we show we can learn reward functions that properly generalize to new settings.


Stochastics ◽  
2021 ◽  
pp. 1-17
Author(s):  
Arnab Bhabak ◽  
Chandan Pal ◽  
Subhamay Saha

Author(s):  
Hao Chen ◽  
Chang Wang ◽  
Jian Huang ◽  
Jianxing Gong
Keyword(s):  

Mathematics ◽  
2021 ◽  
Vol 9 (4) ◽  
pp. 321
Author(s):  
Julio B. Clempner ◽  
Alexander S. Poznyak

A theme that become common knowledge of the literature is the difficulty of developing a mechanism that is compatible with individual incentives that simultaneously result in efficient decisions that maximize the total reward. In this paper, we suggest an analytical method for computing a mechanism design. This problem is explored in the context of a framework, in which the players follow an average utility in a non-cooperative Markov game with incomplete state information. All of the Nash equilibria are approximated in a sequential process. We describe a method for the derivative of the player’s equilibrium that instruments the design of the mechanism. In addition, it showed the convergence and rate of convergence of the proposed method. For computing the mechanism, we consider an extension of the Markov model for which it is introduced a new variable that represents the product of the mechanism design and the joint strategy. We derive formulas to recover the variables of interest: mechanisms, strategy, and distribution vector. The mechanism design and equilibrium strategies computation differ from those in previous literature. A numerical example presents the usefulness and effectiveness of the proposed method.


Sign in / Sign up

Export Citation Format

Share Document