1997 | OriginalPaper | Chapter
Optimal control problems with continuous value functions: unrestricted state space
Authors : Martino Bardi, Italo Capuzzo-Dolcetta
Published in: Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations
Publisher: Birkhäuser Boston
Included in: Professional Book Archive
Activate our intelligent search to find suitable subject content or patents.
Select sections of text to find matching patents with Artificial Intelligence. powered by
Select sections of text to find additional relevant content using AI-assisted search. powered by
In this Chapter we consider several optimal control problems whose value function is defined and continuous on the whole space ℝN. This setting is suitable for those problems where no a priori constraint is imposed on the state of the control system. For all the problems considered we establish the Dynamic Programming Principle and derive from it the appropriate Hamilton-Jacobi-Bellman equation for the value function. This allows us to apply the theory of Chapter II, and some extensions of it, to prove that the value function can in fact be characterized as the unique viscosity solution of the corresponding Hamilton-Jacobi-Bellman equation.