scholarly journals Computing Nash Equilibria in Multiplayer DAG-Structured Stochastic Games with Persistent Imperfect Information

2021 ◽  
pp. 3-16
Author(s):  
Sam Ganzfried
Author(s):  
Krishnendu Chatterjee ◽  
Rupak Majumdar ◽  
Marcin Jurdziński

1999 ◽  
Vol 01 (02) ◽  
pp. 149-158 ◽  
Author(s):  
VICTOR DOMANSKY

Controlled by several agents, multistage processes of resource allocation between production and consumption with random production functions are considered as non-cooperative dynamic stochastic games. For these games, the Nash Equilibria are constructed satisfying the criteria of maximisation of some kind of "public utility". Both finite and infinite horizons of planning are examined.


Author(s):  
Christian Kroer ◽  
Gabriele Farina ◽  
Tuomas Sandholm

Nash equilibrium is a popular solution concept for solving imperfect-information games in practice. However, it has a major drawback: it does not preclude suboptimal play in branches of the game tree that are not reached in equilibrium. Equilibrium refinements can mend this issue, but have experienced little practical adoption. This is largely due to a lack of scalable algorithms.Sparse iterative methods, in particular first-order methods, are known to be among the most effective algorithms for computing Nash equilibria in large-scale two-player zero-sum extensive-form games. In this paper, we provide, to our knowledge, the first extension of these methods to equilibrium refinements. We develop a smoothing approach for behavioral perturbations of the convex polytope that encompasses the strategy spaces of players in an extensive-form game. This enables one to compute an approximate variant of extensive-form perfect equilibria. Experiments show that our smoothing approach leads to solutions with dramatically stronger strategies at information sets that are reached with low probability in approximate Nash equilibria, while retaining the overall convergence rate associated with fast algorithms for Nash equilibrium. This has benefits both in approximate equilibrium finding (such approximation is necessary in practice in large games) where some probabilities are low while possibly heading toward zero in the limit, and exact equilibrium computation where the low probabilities are actually zero.


2020 ◽  
Vol 13 ◽  
pp. 304-323
Author(s):  
Dmitrii Lozovanu ◽  
◽  
Stefan Pickl ◽  

We consider infinite n-person stochastic games with limiting average payoffs criteria for the players. The main results of the paper are concerned with the existence of stationary Nash equilibria and determining the optimal strategies of the players in the games with finite state and action spaces. We present conditions for the existence of stationary Nash equilibria in the considered games and propose an approach for determining the optimal stationary strategies of the players if such strategies exist.


Author(s):  
Matteo Basei ◽  
Haoyang Cao ◽  
Xin Guo

We consider a general class of nonzero-sum N-player stochastic games with impulse controls, where players control the underlying dynamics with discrete interventions. We adopt a verification approach and provide sufficient conditions for the Nash equilibria (NEs) of the game. We then consider the limiting situation when N goes to infinity, that is, a suitable mean-field game (MFG) with impulse controls. We show that under appropriate technical conditions, there exists a unique NE solution to the MFG, which is an ϵ-NE approximation to the N-player game, with [Formula: see text]. As an example, we analyze in detail a class of two-player stochastic games which extends the classical cash management problem to the game setting. In particular, we present numerical analysis for the cases of the single player, the two-player game, and the MFG, showing the impact of competition on the player’s optimal strategy, with sensitivity analysis of the model parameters.


Author(s):  
Yue Guan ◽  
Qifan Zhang ◽  
Panagiotis Tsiotras

We explore the use of policy approximations to reduce the computational cost of learning Nash equilibria in zero-sum stochastic games. We propose a new Q-learning type algorithm that uses a sequence of entropy-regularized soft policies to approximate the Nash policy during the Q-function updates. We prove that under certain conditions, by updating the entropy regularization, the algorithm converges to a Nash equilibrium. We also demonstrate the proposed algorithm's ability to transfer previous training experiences, enabling the agents to adapt quickly to new environments. We provide a dynamic hyper-parameter scheduling scheme to further expedite convergence. Empirical results applied to a number of stochastic games verify that the proposed algorithm converges to the Nash equilibrium, while exhibiting a major speed-up over existing algorithms.


Automatica ◽  
2012 ◽  
Vol 48 (11) ◽  
pp. 2923-2930 ◽  
Author(s):  
H.L. Prasad ◽  
S. Bhatnagar

Sign in / Sign up

Export Citation Format

Share Document