Skip to main content

2020 | Buch

Zero-Sum Discrete-Time Markov Games with Unknown Disturbance Distribution

Discounted and Average Criteria

insite
SUCHEN

Über dieses Buch

This SpringerBrief deals with a class of discrete-time zero-sum Markov games with Borel state and action spaces, and possibly unbounded payoffs, under discounted and average criteria, whose state process evolves according to a stochastic difference equation. The corresponding disturbance process is an observable sequence of independent and identically distributed random variables with unknown distribution for both players. Unlike the standard case, the game is played over an infinite horizon evolving as follows. At each stage, once the players have observed the state of the game, and before choosing the actions, players 1 and 2 implement a statistical estimation process to obtain estimates of the unknown distribution. Then, independently, the players adapt their decisions to such estimators to select their actions and construct their strategies. This book presents a systematic analysis on recent developments in this kind of games. Specifically, the theoretical foundations on the procedures combining statistical estimation and control techniques for the construction of strategies of the players are introduced, with illustrative examples. In this sense, the book is an essential reference for theoretical and applied researchers in the fields of stochastic control and game theory, and their applications.

Inhaltsverzeichnis

Frontmatter
Chapter 1. Zero-Sum Markov Games
Abstract
In this chapter we present the class of zero-sum Markov games we are interested in. We first introduce the game model which is a collection of objects describing the evolution in time of the games. In addition, in order to define the corresponding game problem, the concept of strategies for players is given along with the optimality criteria that will be analyzed in the next chapters.
J. Adolfo Minjárez-Sosa
Chapter 2. Discounted Optimality Criterion
Abstract
We consider the game model
$$\displaystyle \mathscr {G}\mathscr {M}:=(X,A,B,{\mathbb {K}}_{A},{\mathbb {K}}_{B},Q,r) $$
introduced in (1.​1). The problems we are concerned with in this chapter are those related to the discounted case, which are summarized as follows.
J. Adolfo Minjárez-Sosa
Chapter 3. Average Payoff Criterion
Abstract
We are now interested in analyzing the average payoff game, that is, the game in (1.​3) with the long-run expected average payoff in (1.13):
$$\displaystyle J(x,\pi ^{1},\pi ^{2}):=\liminf \limits _{n\rightarrow \infty }\frac {1}{n} E_{x}^{\pi ^{1},\pi ^{2}}\sum _{t=0}^{n-1}r(x_{t},a_{t},b_{t}) {} $$
J. Adolfo Minjárez-Sosa
Chapter 4. Empirical Approximation-Estimation Algorithms in Markov Games
Abstract
This chapter proposes an empirical approximation-estimation algorithm in difference equation game models (see Sect. 1.​1.​1) whose evolution is given by
$$\displaystyle x_{t+1}=F(x_{t},a_{t},b_{t},\xi _{t}), \ t\in \mathbb {N}_{0}, {} $$
where {ξ t} is a sequence of observable i.i.d. random variables defined on a probability space \(\left ( \varOmega ,\mathscr {F},P\right ) ,\) taking values in an arbitrary Borel space S, with common unknown distribution \(\theta \in \mathbb {P}(S).\)
J. Adolfo Minjárez-Sosa
Chapter 5. Difference-Equation Games: Examples
Abstract
In this chapter we introduce several examples to show the relevance of the estimation and control procedures analyzed throughout the book.
J. Adolfo Minjárez-Sosa
Backmatter
Metadaten
Titel
Zero-Sum Discrete-Time Markov Games with Unknown Disturbance Distribution
verfasst von
J. Adolfo Minjárez-Sosa
Copyright-Jahr
2020
Electronic ISBN
978-3-030-35720-7
Print ISBN
978-3-030-35719-1
DOI
https://doi.org/10.1007/978-3-030-35720-7