Discrete-time Stochastic Systems


Book Description

This comprehensive introduction to the estimation and control of dynamic stochastic systems provides complete derivations of key results. The second edition includes improved and updated material, and a new presentation of polynomial control and new derivation of linear-quadratic-Gaussian control.




Stochastic Control in Discrete and Continuous Time


Book Description

This book contains an introduction to three topics in stochastic control: discrete time stochastic control, i. e. , stochastic dynamic programming (Chapter 1), piecewise - terministic control problems (Chapter 3), and control of Ito diffusions (Chapter 4). The chapters include treatments of optimal stopping problems. An Appendix - calls material from elementary probability theory and gives heuristic explanations of certain more advanced tools in probability theory. The book will hopefully be of interest to students in several ?elds: economics, engineering, operations research, ?nance, business, mathematics. In economics and business administration, graduate students should readily be able to read it, and the mathematical level can be suitable for advanced undergraduates in mathem- ics and science. The prerequisites for reading the book are only a calculus course and a course in elementary probability. (Certain technical comments may demand a slightly better background. ) As this book perhaps (and hopefully) will be read by readers with widely diff- ing backgrounds, some general advice may be useful: Don’t be put off if paragraphs, comments, or remarks contain material of a seemingly more technical nature that you don’t understand. Just skip such material and continue reading, it will surely not be needed in order to understand the main ideas and results. The presentation avoids the use of measure theory.




Stochastic Optimal Control


Book Description




Control and System Theory of Discrete-Time Stochastic Systems


Book Description

This book helps students, researchers, and practicing engineers to understand the theoretical framework of control and system theory for discrete-time stochastic systems so that they can then apply its principles to their own stochastic control systems and to the solution of control, filtering, and realization problems for such systems. Applications of the theory in the book include the control of ships, shock absorbers, traffic and communications networks, and power systems with fluctuating power flows. The focus of the book is a stochastic control system defined for a spectrum of probability distributions including Bernoulli, finite, Poisson, beta, gamma, and Gaussian distributions. The concepts of observability and controllability of a stochastic control system are defined and characterized. Each output process considered is, with respect to conditions, represented by a stochastic system called a stochastic realization. The existence of a control law is related to stochastic controllability while the existence of a filter system is related to stochastic observability. Stochastic control with partial observations is based on the existence of a stochastic realization of the filtration of the observed process.​




Optimal Control of Discrete Time Stochastic Systems


Book Description

Introduction and formulation of the model; Estimation; Statistics sufficient for control; General theory of optimality; Selection classes; Quadratic loss; An absolute value loss function.




Mathematical Methods in Robust Control of Discrete-Time Linear Stochastic Systems


Book Description

In this monograph the authors develop a theory for the robust control of discrete-time stochastic systems, subjected to both independent random perturbations and to Markov chains. Such systems are widely used to provide mathematical models for real processes in fields such as aerospace engineering, communications, manufacturing, finance and economy. The theory is a continuation of the authors’ work presented in their previous book entitled "Mathematical Methods in Robust Control of Linear Stochastic Systems" published by Springer in 2006. Key features: - Provides a common unifying framework for discrete-time stochastic systems corrupted with both independent random perturbations and with Markovian jumps which are usually treated separately in the control literature; - Covers preliminary material on probability theory, independent random variables, conditional expectation and Markov chains; - Proposes new numerical algorithms to solve coupled matrix algebraic Riccati equations; - Leads the reader in a natural way to the original results through a systematic presentation; - Presents new theoretical results with detailed numerical examples. The monograph is geared to researchers and graduate students in advanced control engineering, applied mathematics, mathematical systems theory and finance. It is also accessible to undergraduate students with a fundamental knowledge in the theory of stochastic systems.







Dynamic Programming and Optimal Control


Book Description

"The leading and most up-to-date textbook on the far-ranging algorithmic methododogy of Dynamic Programming, which can be used for optimal control, Markovian decision problems, planning and sequential decision making under uncertainty, and discrete/combinatorial optimization. The treatment focuses on basic unifying themes, and conceptual foundations. It illustrates the versatility, power, and generality of the method with many examples and applications from engineering, operations research, and other fields. It also addresses extensively the practical application of the methodology, possibly through the use of approximations, and provides an extensive treatment of the far-reaching methodology of Neuro-Dynamic Programming/Reinforcement Learning. The first volume is oriented towards modeling, conceptualization, and finite-horizon problems, but also includes a substantive introduction to infinite horizon problems that is suitable for classroom use. The second volume is oriented towards mathematical analysis and computation, treats infinite horizon problems extensively, and provides an up-to-date account of approximate large-scale dynamic programming and reinforcement learning. The text contains many illustrations, worked-out examples, and exercises."--Publisher's website.




Linear Stochastic Control Systems


Book Description

Linear Stochastic Control Systems presents a thorough description of the mathematical theory and fundamental principles of linear stochastic control systems. Both continuous-time and discrete-time systems are thoroughly covered. Reviews of the modern probability and random processes theories and the Itô stochastic differential equations are provided. Discrete-time stochastic systems theory, optimal estimation and Kalman filtering, and optimal stochastic control theory are studied in detail. A modern treatment of these same topics for continuous-time stochastic control systems is included. The text is written in an easy-to-understand style, and the reader needs only to have a background of elementary real analysis and linear deterministic systems theory to comprehend the subject matter. This graduate textbook is also suitable for self-study, professional training, and as a handy research reference. Linear Stochastic Control Systems is self-contained and provides a step-by-step development of the theory, with many illustrative examples, exercises, and engineering applications.




Discrete-Time Markov Jump Linear Systems


Book Description

This will be the most up-to-date book in the area (the closest competition was published in 1990) This book takes a new slant and is in discrete rather than continuous time