On sets of occupational measures generated by a deterministic control system on an infinite time horizon

https://doi.org/10.1016/j.na.2013.03.015Get rights and content

Abstract

We give a representation for the closed convex hull of the set of discounted occupational measures generated by control-state trajectories of a deterministic control system. We also investigate the limit behavior of the latter when the discount factor tends to zero and compare it with the limit behavior of the long run time average occupational measures set. The novelty of our results is in that we allow the control set dependence on the state variables that make the results to be applicable to differential inclusions.

Section snippets

Introduction and preliminaries

It is well known that nonlinear optimal control problems can be equivalently reformulated as infinite dimensional linear programming problems considered on spaces of occupational measures generated by control-state trajectories. Having many attractive features and being applicable in both stochastic and deterministic settings, the linear programming (LP) based approaches to optimal control problems have been intensively studied in the literature. Important results justifying the use of LP

A representation of the set of discounted occupational measures

Lemma 2.1

The following inclusion is validcōΓdisr(C,y0)W(C,y0),where cō stands for the closed convex hull, andW(C,y0)=def{γP(K):K(ϕ(y)Tf(u,y)+C(ϕ(y0)ϕ(y)))γ(du,dy)=0ϕC1}.

Proof

Take arbitrary γΓdisr(C,y0). By definition, there exists a relaxed y0-admissible pair (μ(),y()) such that γ is the discounted occupational measure generated by this pair. Using the fact that (11) is valid for any continuous function h(u,y), one can obtain Kϕ(y)Tf(u,y)γ(du,dy)=C0eCt(ϕ(y(t)))T(U(y(t))f(u,y(t))μ(t,du))dt=C

Proof of Theorem 2.2

We will divide the proof into four steps.

(i) Auxiliary relaxed admissible pairs. Let the multivalued function F() be defined by the equation F(y)=def{v:v=U(y)f(u,y)μ(du),μP(M),supp(μ)U(y)}yY. It is easy to verify that F() is upper semicontinuous. Hence, its graph Graph(F)=def{(v,y):vF(y),yY} is compact.

Let D and Q be closed balls in Rm such that YD and F(y)QyYGraph(F)Q×D.

Let ν(t,dv):[0,)P(Q) be measurable and let y(t) satisfy the equation y(t)=Qvν(t,dv)for a.e.t>0;y(0)=y0.

Proof of Lemma 2.4

To prove (22), it is enough to show that (28) implies (32). Note that from (28) (and from (25)) it follows that, for any sequence Ci0, there exist a sequence of y0iY and a sequence of y0i-admissible pairs (ui(),yi()) such that limiζi=0,ζi=defCi0+eCitg(ui(t),yi(t))dtG. From Lemma 3.5(ii) in [28] it follows that there exists a sequence Si, i=1,2,, such that SiKCi (K>0 being a constant) and such that 1Si0Sig(ui(t),yi(t))dtG+ζi+Ciinfy0YΘSi(y0)ΘSi(y0i)G+ζi+Cilim¯Sinfy0YΘS(y0)

Acknowledgments

The work of the first author was partially funded by the Australian Research Council Discovery-Project Grants DP0664330, DP120100532, and DP130104432 and by the Linkage International Grant LX0560049. The second author was partially supported by project SADCO, FP7-PEOPLE-2010-ITN, No. 264735 and he was also supported partially by the French National Research AgencyANR-10-BLAN 0112.

References (29)

  • M. Quincampoix et al.

    The problem of optimal control with reflection studied through a linear optimization problem stated on occupational measures

    Nonlinear Anal.

    (2010)
  • H. Frankowska et al.

    Filippov’s and Filippov–Wazewski’s theorems on closed domains

    Journal of Differential Equations

    (2000)
  • L. Grüne

    On the relation between discounted and average optimal value functions

    Journal of Differential Equations

    (1998)
  • A.G. Bhatt et al.

    Occupation measures for controlled markov processes: characterization and optimality

    Annals of Probability

    (1996)
  • W.H. Fleming et al.

    Convex duality approach to the optimal control of diffusions

    SIAM Journal on Control and Optimization

    (1989)
  • T.G. Kurtz et al.

    Existence of Markov controls and characterization of optimal markov controls

    SIAM Journal on Control and Optimization

    (1998)
  • R.H. Stockbridge

    Time-Average Control of a Martingale Problem. Existence of a Stationary Solution

    Annals of Probability

    (1990)
  • R.H. Stockbridge, Time-average control of a martingale problem: a linear programming formulation, Annals of...
  • V. Borkar et al.

    Ergodic control for constrained diffusions: characterization using HJB equations. (English summary)

    SIAM Journal on Control and Optimization

    (2004/05)
  • R. Buckdahn et al.

    Stochastic optimal control and linear programming approach

    Applied Mathematics and Optimization

    (2011)
  • F. Dufour et al.

    On the existence of strict optimal controls for constrained, controlled Markov processes in continuous-time

    Stochastics

    (2012)
  • D. Goreac et al.

    A note on linearization methods and dynamic programming principles for stochastic discontinuous control problems

    Electronic Communications in Probability

    (2012)
  • D. Hernandez-Hernandez et al.

    The linear programming approach to deterministic optimal control problems

    Applicationes Mathematicae

    (1996)
  • O. Hernandez-Lerma et al.

    Markov Chains and Invariant Probabilities

    (2003)
  • Cited by (31)

    View all citing articles on Scopus
    View full text