REINFORCEMENT LEARNING IN MARKOVIAN EVOLUTIONARY GAMES

2002 ◽  
Vol 05 (01) ◽  
pp. 55-72 ◽  
Author(s):  
V. S. BORKAR

A population of agents plays a stochastic dynamic game wherein there is an underlying state process with a Markovian dynamics that also affects their costs. A learning mechanism is proposed which takes into account intertemporal effects and incorporates an explicit process of expectation formation. The agents use this scheme to update their mixed strategies incrementally. The asymptotic behavior of this scheme is captured by an associated ordinary differential equation. Both the formulation and the analysis of the scheme draw upon the theory of reinforcement learning in artificial intelligence.

2018 ◽  
Vol 23 (4) ◽  
pp. 452-477
Author(s):  
Xiao-Bing Zhang ◽  
Magnus Hennlock

AbstractThis paper investigates the benefits of international cooperation under uncertainty about global warming through a stochastic dynamic game. We analyze the benefits of cooperation both for the case of symmetric and asymmetric players. It is shown that the players’ combined expected payoffs decrease as climate uncertainty becomes larger, whether or not they cooperate. However, the benefits from cooperation increase with climate uncertainty. In other words, it is more important to cooperate when facing higher uncertainty. At the same time, more transfers will be needed to ensure stable cooperation among asymmetric players.


2020 ◽  
Vol 99 (4) ◽  
pp. 3301-3312 ◽  
Author(s):  
Si-Ping Zhang ◽  
Ji-Qiang Zhang ◽  
Li Chen ◽  
Xu-Dong Liu

2020 ◽  
Vol 7 (8) ◽  
pp. 200891 ◽  
Author(s):  
Hiromu Ito ◽  
Jun Tanimoto

Game theory has been extensively applied to elucidate the evolutionary mechanism of cooperative behaviour. Dilemmas in game theory are important elements that disturb the promotion of cooperation. An important question is how to escape from dilemmas. Recently, a dynamic utility function (DUF) that considers an individual's current status (wealth) and that can be applied to game theory was developed. The DUF is different from the famous five reciprocity mechanisms called Nowak's five rules. Under the DUF, cooperation is promoted by poor players in the chicken game, with no changes in the prisoner's dilemma and stag-hunt games. In this paper, by comparing the strengths of the two dilemmas, we show that the DUF is a novel reciprocity mechanism (sixth rule) that differs from Nowak's five rules. We also show the difference in dilemma relaxation between dynamic game theory and (traditional) static game theory when the DUF and one of the five rules are combined. Our results indicate that poor players unequivocally promote cooperation in any dynamic game. Unlike conventional rules that have to be brought into game settings, this sixth rule is universally (canonical form) applicable to any game because all repeated/evolutionary games are dynamic in principle.


1996 ◽  
Vol 51 (3) ◽  
pp. 501-518 ◽  
Author(s):  
JEFFREY R. LUCAS ◽  
RICHARD D. HOWARD ◽  
JOSEPH G. PALMER

Sign in / Sign up

Export Citation Format

Share Document