Skip to main content
Log in

Continuity of the Value of Competitive Markov Decision Processes

  • Published:
Journal of Theoretical Probability Aims and scope Submit manuscript

Abstract

We provide a bound for the variation of the function that assigns to every competitive Markov decision process and every discount factor its discounted value. This bound implies that the undiscounted value of a competitive Markov decision process is continuous in the relative interior of the space of transition rules.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Amir, R. (1987). Sequential Games of Resource Extraction: Existence of Nash Equilibrium, Cowles Foundation D.P. #825.

  2. Arapostathis, A., Borkar, V. S., Fernández-Gaucherand, E.,Ghosh, M. K., and Marcus, S. I. (1993). Discrete-time controlled Markov processes with average cost criterion: A survey. SIAM J. Control Optim. 31,282–344.

    Google Scholar 

  3. Bewley, T., and Kohlberg, E. (1976). The asymptotic theory of stochastic games. Math. Oper. Res. 1, 197–208.

    Google Scholar 

  4. Catoni, O. (1999). Simulated Annealing Algorithms and Markov Chains with Rare Transitions, Séminaire de Probabilités, XXXIII, 69–119, Lecture Notes in Mathematics, 1709, Springer, Berlin.

    Google Scholar 

  5. Filar, J. A. (1985). Player aggregation in the traveling inspector model, IEEE Trans. Automatic Control AC-30, 723–729.

    Google Scholar 

  6. Filar, J. A., and Vrieze, K. (1997). Competitive Markov Decision Processes, Springer-Verlag, New York.

    Google Scholar 

  7. Freidlin, M. I., and Wentzell, A. D. (1984). Random Perturbations of Dynamical Systems, Springer-Verlag, Berlin.

    Google Scholar 

  8. Levhari, D., and Mirman, L. (1980). The great fish war: An example using a dynamic Cournot–Nash solution. Bell J. Econ. 11, 322–334.

    Google Scholar 

  9. Mertens, J. F., and Neyman, A. (1981). Stochastic games. Int. J. Game theory 10, 53–66.

    Google Scholar 

  10. Milman, E. (2002). The semi-algebraic theory of stochastic games. Math. Oper. Res. 27, 401–418.

    Google Scholar 

  11. Schweizer, P. J. (1968). Perturbation theory and finite Markov chains. J. Applied Probab. 5, 401–413.

    Google Scholar 

  12. Shapley, L. S. (1953). Stochastic games. Proc. Nat. Acad. Sci. U.S.A. 39, 1095–1100.

    Google Scholar 

  13. Sorin, S. (2002). A First Course on Zero-Sum Repeated Games, Mathématiques et Applications, Vol. 37, Springer-Verlag, Berlin.

    Google Scholar 

  14. Winston, W. (1978). A stochastic game model of a weapons development competition. SIAM J. Control Optim. 16, 411–419.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Solan, E. Continuity of the Value of Competitive Markov Decision Processes. Journal of Theoretical Probability 16, 831–845 (2003). https://doi.org/10.1023/B:JOTP.0000011995.28536.ef

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/B:JOTP.0000011995.28536.ef

Navigation