확률 게임
Stochastic game게임 이론에서 로이드 샤플리가 1950년대 초 선보인 확률형 게임은 한 명 이상의 플레이어가 펼치는 확률적 전환이 반복되는 게임이다.[1] 그 게임은 일련의 단계로 진행된다. 각 스테이지가 시작될 때 경기는 어떤 상태에 있다. 플레이어는 행동을 선택하고 각 플레이어는 현재 상태와 선택한 행동에 따라 보상을 받는다. 그런 다음 게임은 이전 상태와 플레이어가 선택한 행동에 따라 분포가 달라지는 새로운 무작위 상태로 이동한다. 절차는 새로운 상태에서 반복되며 극은 한정된 또는 무한정 많은 스테이지에 대해 계속된다. 플레이어에 대한 총 보수는 스테이지 보상의 할인된 합계 또는 스테이지 보상의 평균보다 낮은 한도로 간주되는 경우가 많다.
확률형 게임은 마르코프 의사결정 프로세스를 여러 상호 작용하는 의사결정자에게 일반화하며, 전략형 게임은 플레이어들의 선택에 따라 환경이 변화하는 역동적인 상황에 맞춰 일반화한다.[2]
2인용 게임
지시된 그래프의 확률형 2인용 게임은 알려지지 않은 (수리적) 환경에서 작동하는 이산 시스템의 모델링 및 분석에 널리 사용된다. 시스템과 그 환경의 가능한 구성은 정점으로 표시되며, 전환은 시스템, 그 환경 또는 "자연"의 작용에 해당한다. 그러면 시스템의 실행은 그래프의 무한 경로에 해당된다. 따라서 한 선수(시스템)는 '선행'의 확률 극대화를, 다른 선수(환경)는 정반대를 지향하는 적대적 목표를 가진 두 선수로 볼 수 있다.
많은 경우, 이 확률의 평형값이 존재하지만, 두 선수 모두를 위한 최적의 전략은 존재하지 않을 수 있다.
We introduce basic concepts and algorithmic questions studied in this area, and we mention some long-standing open problems. Then, we mention selected recent results.
Theory
The ingredients of a stochastic game are: a finite set of players ; a state space (either a finite set or a measurable space ); for each player , an action set (either a finite set or a measurable space ); a transition probability from , where is the action profiles, to , where is the probability that the next state is in given the current state and the current action profile ; and a payoff function from to , where the -th coordinate of , , is the payoff to player as a function of the state and the action profile .
게임은 어떤 초기 m 에서 시작된다 t{\ 플레이어는 m 동시에 동작 를 선택하고 =( ) }{을 관찰한다., and then nature selects according to the probability . A play of the stochastic game, 은는) g , … 여기서 = s ) 의 지급 스트림을 정의한다
The discounted game with discount factor () is the game where the payoff to player is . The -stage game is the game where the payoff to player is .
The value , respectively , of a two-person zero-sum stochastic game , respectively , with finitely many states and actions exists, and Truman Bewley and Elon Kohlberg (1976) proved that converges to a limit as goes to infinity and that converges to the same limit as goes to .
The "undiscounted" game is the game where the payoff to player is the "limit" of the averages of the stage payoffs. Some precautions are needed in defining the value of a two-person zero-sum and in defining equilibrium payoffs of a non-zero-sum . The uniform value of a two-person zero-sum stochastic game exists if for every there is a positive integer and a strategy pair of player 1 and of player 2 such that for every and and every the expectation of with respect to the probability on plays defined by and is at least , and the expectation of with respect to the probability on plays defined by and is at most . Jean-François Mertens and Abraham Neyman (1981) proved that every two-person zero-sum stochastic game with finitely many states and actions has a uniform value.[3]
If there is a finite number of players and the action sets and the set of states are finite, then a stochastic game with a finite number of stages always has a Nash equilibrium. The same is true for a game with infinitely many stages if the total payoff is the discounted sum.
The non-zero-sum stochastic game has a uniform equilibrium payoff if for every there is a positive integer and a strategy profile such that for every unilateral deviation by a player , i.e., a strategy profile with for all , and every the expectation of with respect to the probability on plays defined by is at least , and the expectation of with respect to the probability on plays defined by is at most . Nicolas Vieille has shown that all two-person stochastic games with finite state and action spaces have a uniform equilibrium payoff.[4]
The non-zero-sum stochastic game has a limiting-average equilibrium payoff if for every there is a strategy profile such that for every unilateral deviation by a player , the expectation of the limit inferior of the averages of the stage payoffs with respect to the probability on plays defined by is at least , and the expectation of the limit superior of the averages of the stage payoffs with respect to the probability on plays defined by is at most . Jean-François Mertens and Abraham Neyman (1981) proves that every two-person zero-sum stochastic game with finitely many states and actions has a limiting-average value,[3] and Nicolas Vieille has shown that all two-person stochastic games with finite state and action spaces have a limiting-average equilibrium payoff.[4] In particular, these results imply that these games have a value and an approximate equilibrium payoff, called the liminf-average (respectively, the limsup-average) equilibrium payoff, when the total payoff is the limit inferior (or the limit superior) of the averages of the stage payoffs.
Whether every stochastic game with finitely many players, states, and actions, has a uniform equilibrium payoff, or a limiting-average equilibrium payoff, or even a liminf-average equilibrium payoff, is a challenging open question.
A Markov perfect equilibrium is a refinement of the concept of sub-game perfect Nash equilibrium to stochastic games.
Stochastic games have been combined with Bayesian games to model uncertainty over player strategies.[5] The resulting "stochastic Bayesian game" model is solved via a recursive combination of the Bayesian Nash equilibrium equation and the Bellman optimality equation.
Applications
Stochastic games have applications in economics, evolutionary biology and computer networks.[6][7] They are generalizations of repeated games which correspond to the special case where there is only one state.
See also
Notes
- ^ Shapley, L. S. (1953). "Stochastic games". PNAS. 39 (10): 1095–1100. Bibcode:1953PNAS...39.1095S. doi:10.1073/pnas.39.10.1095. PMC 1063912. PMID 16589380.
- ^ Solan, Eilon; Vieille, Nicolas (2015). "Stochastic Games". PNAS. 112 (45): 13743–13746. doi:10.1073/pnas.1513508112. PMC 4653174. PMID 26556883.
- ^ a b Mertens, J. F. & Neyman, A. (1981). "Stochastic Games". International Journal of Game Theory. 10 (2): 53–66. doi:10.1007/BF01769259. S2CID 189830419.
- ^ a b Vieille, N. (2002). "Stochastic games: Recent results". Handbook of Game Theory. Amsterdam: Elsevier Science. pp. 1833–1850. ISBN 0-444-88098-4.
- ^ Albrecht, Stefano; Crandall, Jacob; Ramamoorthy, Subramanian (2016). "Belief and Truth in Hypothesised Behaviours". Artificial Intelligence. 235: 63–94. arXiv:1507.07688. doi:10.1016/j.artint.2016.02.004. S2CID 2599762.
- ^ Constrained Stochastic Games in Wireless Networks by E.Altman, K.Avratchenkov, N.Bonneau, M.Debbah, R.El-Azouzi, D.S.Menasche
- ^ Djehiche, Boualem; Tcheukam, Alain; Tembine, Hamidou (2017-09-27). "Mean-Field-Type Games in Engineering". AIMS Electronics and Electrical Engineering. 1: 18–73. arXiv:1605.03281. doi:10.3934/ElectrEng.2017.1.18. S2CID 16055840.
Further reading
- Filar, J. & Vrieze, K. (1997). Competitive Markov Decision Processes. Springer-Verlag. ISBN 0-387-94805-8.
- Neyman, A. & Sorin, S. (2003). Stochastic Games and Applications. Dordrecht: Kluwer Academic Press. ISBN 1-4020-1492-9.
- Yoav Shoham; Kevin Leyton-Brown (2009). Multiagent systems: algorithmic, game-theoretic, and logical foundations. Cambridge University Press. pp. 153–156. ISBN 978-0-521-89943-7. (suitable for undergraduates; main results, no proofs)