This book is an introduction to optimal stochastic control for
continuous time Markov processes and the theory of viscosity solutions.
The authors approach stochastic control problems by the method of
dynamic programming. The text covers dynamic programming for
deterministic optimal control problems, as well as to the corresponding
theory of viscosity solutions. New chapters introduce the role of
stochastic optimal control in portfolio optimization and in pricing
derivatives in incomplete markets and two-controller, zero-sum
differential games. Also covered are controlled Markov diffusions and
viscosity solutions of Hamilton-Jacobi-Bellman equations. The authors
use illustrative examples and selective material to connect stochastic
control theory with other mathematical areas (e.g. large deviations
theory) and with applications to engineering, physics, management, and
finance.