全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...
Game Theory  2013 

A Necessary Condition for Nash Equilibrium in Two-Person Zero-Sum Constrained Stochastic Games

DOI: 10.1155/2013/290427

Full-Text   Cite this paper   Add to My Lib

Abstract:

We provide a necessary condition that a constrained Nash-equilibrium (CNE) policy pair satisfies in two-person zero-sum constrained stochastic discounted-payoff games and discuss a general method of approximating CNE based on the condition. 1. Introduction Altman and Shwartz [1] established a sufficient condition for the existence of a stationary Markovian constrained Nash equilibrium (CNE) policy pair in a general model of finite two-person zero-sum constrained stochastic games and Alvarez-Mena and Hernández-Lerma [2] extended the result for infinite state and action spaces. Even though a few computational studies exist for average-payoff models with additional simplifying assumptions (see, e.g., [3–5]), there seems to be no work providing a meaningful necessary condition for CNE or any general approximation scheme for CNE within the general discounting-cost model. This brief paper establishes a necessary condition that a CNE policy pair satisfies by a novel characterization of the set of all feasible policies of one player when the other player's policy is fixed. This is done by identifying feasible mixed actions of one player at a current state when the expected total discounted constraint cost from each reachable next state is given by a value function defined over the state space. The necessary condition provides a general method of testing whether a given policy pair is a CNE policy pair and can induce a general approximation scheme for CNE. 2. Preliminaries Consider a two-person zero-sum Markov game (MG) [6] , where is a finite state set, and and are nonempty finite pure-action sets for the minimizer and the maximizer, respectively, at in with and . We denote the mixed action sets at in over and to be and , respectively, with and . Once in and in are simultaneously taken at in by the minimizer and the maximizer, respectively (with the complete knowledge of the state but without knowing each other's current action being taken), makes a transition to a next state by the probability given as Here denotes the probability of selecting , similar to , and denotes the probability of moving from to by and . Then the minimizer obtains an expected cost of given by where in is a payoff to the minimizer (the negative of this will be incurred to the maximizer). We define a stationary Markovian policy?? of the minimizer as a function with for all and denote to be the set of all possible such policies. A policy is similarly defined for the maximizer with , and we denote to be the set of all possible such policies. Define the objective value of in and in with an

References

[1]  E. Altman and A. Shwartz, “Constrained Markov games: Nash equilibria,” in Annals of the International Society of Dynamic Games, V. Gaitsgory, J. Filar, and K. Mizukami, Eds., vol. 5, pp. 303–323, Birkh?auser, Boston, Mass, USA, 2000.
[2]  J. Alvarez-Mena and O. Hernández-Lerma, “Existence of nash equilibria for constrained stochastic games,” Mathematical Methods of Operations Research, vol. 63, no. 2, pp. 261–285, 2006.
[3]  E. Altman, K. Avrachenkov, R. Marquez, and G. Miller, “Zero-sum constrained stochastic games with independent state processes,” Mathematical Methods of Operations Research, vol. 62, no. 3, pp. 375–386, 2005.
[4]  E. Altman, K. Avrachenkov, N. Bonneau, M. Debbah, R. El-Azouzi, and D. S. Menasche, “Constrained cost-coupled stochastic games with independent state processes,” Operations Research Letters, vol. 36, no. 2, pp. 160–164, 2008.
[5]  N. Shimkin, “Stochastic games with average cost constraints,” in Annals of the International Society of Dynamic Games, T. Basar and A. Haurie, Eds., vol. 1 of Advances in Dynamic Games and Applications, Birkh?auser, Boston, Mass, USA, 1994.
[6]  J. Filar and K. Vrieze, Competitive Markov Decision Processes, Springer, New York, NY, USA, 1996.
[7]  C. Alós-Ferrer and A. B. Ania, “Local equilibria in economic games,” Economics Letters, vol. 70, no. 2, pp. 165–173, 2001.
[8]  D. S. Hochbaum, “Approximating clique and biclique problems,” Journal of Algorithms, vol. 29, no. 1, pp. 174–200, 1998.
[9]  M. L. Puterman, Markov Decision Processes: Discrete Stochastic Dynamic Programming, John Wiley & Sons, New York, NY, USA, 1994.

Full-Text

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133