Zero-sum two-person semi-Markov games

1992 ◽  
Vol 29 (1) ◽  
pp. 56-72 ◽  
Author(s):  
Arbind K. Lal ◽  
Sagnik Sinha

Semi-Markov games are investigated under discounted and limiting average payoff criteria. The issue of the existence of the value and a pair of stationary optimal strategies are settled; the optimality equation is studied and under a natural ergodic condition the existence of a solution to the optimality equation is proved for the limiting average case. Semi-Markov games provide useful flexibility in constructing recursive game models. All the work on Markov/semi-Markov decision processes and Markov (stochastic) games can be viewed as special cases of the developments in this paper.

1992 ◽  
Vol 29 (01) ◽  
pp. 56-72 ◽  
Author(s):  
Arbind K. Lal ◽  
Sagnik Sinha

Semi-Markov games are investigated under discounted and limiting average payoff criteria. The issue of the existence of the value and a pair of stationary optimal strategies are settled; the optimality equation is studied and under a natural ergodic condition the existence of a solution to the optimality equation is proved for the limiting average case. Semi-Markov games provide useful flexibility in constructing recursive game models. All the work on Markov/semi-Markov decision processes and Markov (stochastic) games can be viewed as special cases of the developments in this paper.


2001 ◽  
Vol 03 (02n03) ◽  
pp. 253-281 ◽  
Author(s):  
S. R. MOHAN ◽  
S. K. NEOGY ◽  
T. PARTHASARATHY

In this paper, we survey the recent literature on computing the value vector and the associated optimal strategies of the players for special cases of zero-sum stochastic games, or in computing a Nash equilibrium point and the corresponding stationary strategies of the players for special cases of nonzero-sum stochastic games, using finite-step algorithms based on pivoting. Examples of finite-step pivoting algorithms are the various simplex-type algorithms, such as the primal simplex or dual simplex method for solving the linear programming problem or Lemke's or Lemke-Howson's algorithm for solving the linear complementarity problem. Also included are Lemke-type algorithms for solving various generalisations of the linear complementarity problem. The survey also includes a few new results and observations.


2001 ◽  
Vol 54 (2) ◽  
pp. 291-301 ◽  
Author(s):  
Anna Jaśkiewicz ◽  
Andrzej S. Nowak

2017 ◽  
Vol 49 (3) ◽  
pp. 826-849 ◽  
Author(s):  
Prasenjit Mondal

Abstract Zero-sum two-person finite undiscounted (limiting ratio average) semi-Markov games (SMGs) are considered with a general multichain structure. We derive the strategy evaluation equations for stationary strategies of the players. A relation between the payoff in the multichain SMG and that in the associated stochastic game (SG) obtained by a data-transformation is established. We prove that the multichain optimality equations (OEs) for an SMG have a solution if and only if the associated SG has optimal stationary strategies. Though the solution of the OEs may not be optimal for an SMG, we establish the significance of studying the OEs for a multichain SMG. We provide a nice example of SMGs in which one player has no optimal strategy in the stationary class but has an optimal semistationary strategy (that depends only on the initial and current state of the game). For an SMG with absorbing states, we prove that solutions in the game where all players are restricted to semistationary strategies are solutions for the unrestricted game. Finally, we prove the existence of stationary optimal strategies for unichain SMGs and conclude that the unichain condition is equivalent to require that the game satisfies some recurrence/ergodicity/weakly communicating conditions.


2014 ◽  
Vol 1 (1) ◽  
pp. 105-119
Author(s):  
Fernando Luque-Vásquez ◽  
◽  
J. Adolfo Minjárez-Sosa ◽  

2020 ◽  
Vol 22 (02) ◽  
pp. 2040008
Author(s):  
P. Mondal ◽  
S. K. Neogy ◽  
A. Gupta ◽  
D. Ghorui

Zero-sum two-person discounted semi-Markov games with finite state and action spaces are studied where a collection of states having Perfect Information (PI) property is mixed with another collection of states having Additive Reward–Additive Transition and Action Independent Transition Time (AR-AT-AITT) property. For such a PI/AR-AT-AITT mixture class of games, we prove the existence of an optimal pure stationary strategy for each player. We develop a policy improvement algorithm for solving discounted semi-Markov decision processes (one player version of semi-Markov games) and using it we obtain a policy-improvement type algorithm for computing an optimal strategy pair of a PI/AR-AT-AITT mixture semi-Markov game. Finally, we extend our results when the states having PI property are replaced by a subclass of Switching Control (SC) states.


2009 ◽  
Vol 48 (3) ◽  
pp. 1405-1421 ◽  
Author(s):  
J. Adolfo Minjárez-Sosa ◽  
Oscar Vega-Amaya

2015 ◽  
Vol 32 (06) ◽  
pp. 1550043 ◽  
Author(s):  
Prasenjit Mondal

In this paper, zero-sum two-person finite undiscounted (limiting average) semi-Markov games (SMGs) are considered. We prove that the solutions of the game when both players are restricted to semi-Markov strategies are solutions for the original game. In addition, we show that if one player fixes a stationary strategy, then the other player can restrict himself in solving an undiscounted semi-Markov decision process associated with that stationary strategy. The undiscounted SMGs are also studied when the transition probabilities and the transition times are controlled by a fixed player in all states. If such games are unichain, we prove that the value and optimal stationary strategies of the players can be obtained from an optimal solution of a linear programming algorithm. We propose a realistic and generalized traveling inspection model that suitably fits into the class of one player control undiscounted unichain semi-Markov games.


2012 ◽  
Vol 40 (1) ◽  
pp. 56-60 ◽  
Author(s):  
Konstantin Avrachenkov ◽  
Laura Cottatellucci ◽  
Lorenzo Maggi

Sign in / Sign up

Export Citation Format

Share Document