Stochastic Control with Complete Observations on an Infinite Horizon
J.H. van Schuppen (TU Delft - Mathematical Physics)
More Info
expand_more
Abstract
Optimal stochastic control problems with complete observations and on an infinite horizon are considered. Control theory for both the average cost and the discounted cost function is treated. The dynamic programming approach is formulated as a procedure to determine the value and the value function; from the value function, one can derive the optimal control law. Stochastic controllability is in general needed to prove that there exists a control law with a finite average cost in case of positive cost. Special cases treated in depth are: the case of a Gaussian stochastic control system and of a finite stochastic control system.
No files available
Metadata only record. There are no files for this record.