This book is devoted to an investigation of control problems which can
be described by ordinary differential equations and be expressed in
terms of game theoretical notions. In these terms, a strategy is a
control based on the feedback principle which will assure a definite
equality for the controlled process which is subject to uncertain
factors such as a move or a controlling action of the opponent. Game
Theoretical Control Problems contains definitions and formalizations
of differential games, existence for equilibrium and extensive
discussions of optimal strategies. Formal definitions and statements are
accompanied by suitable motivations and discussions of computational
algorithms. The book is addessed to mathematicians, engineers,
economists and other users of control theoretical and game theoretical
notions.