Markov Processes for Stochastic Modeling


Book Description

Markov processes are processes that have limited memory. In particular, their dependence on the past is only through the previous state. They are used to model the behavior of many systems including communications systems, transportation networks, image segmentation and analysis, biological systems and DNA sequence analysis, random atomic motion and diffusion in physics, social mobility, population studies, epidemiology, animal and insect migration, queueing systems, resource management, dams, financial engineering, actuarial science, and decision systems. Covering a wide range of areas of application of Markov processes, this second edition is revised to highlight the most important aspects as well as the most recent trends and applications of Markov processes. The author spent over 16 years in the industry before returning to academia, and he has applied many of the principles covered in this book in multiple research projects. Therefore, this is an applications-oriented book that also includes enough theory to provide a solid ground in the subject for the reader. - Presents both the theory and applications of the different aspects of Markov processes - Includes numerous solved examples as well as detailed diagrams that make it easier to understand the principle being presented - Discusses different applications of hidden Markov models, such as DNA sequence analysis and speech analysis.




An Introduction to Stochastic Modeling


Book Description

An Introduction to Stochastic Modeling provides information pertinent to the standard concepts and methods of stochastic modeling. This book presents the rich diversity of applications of stochastic processes in the sciences. Organized into nine chapters, this book begins with an overview of diverse types of stochastic models, which predicts a set of possible outcomes weighed by their likelihoods or probabilities. This text then provides exercises in the applications of simple stochastic analysis to appropriate problems. Other chapters consider the study of general functions of independent, identically distributed, nonnegative random variables representing the successive intervals between renewals. This book discusses as well the numerous examples of Markov branching processes that arise naturally in various scientific disciplines. The final chapter deals with queueing models, which aid the design process by predicting system performance. This book is a valuable resource for students of engineering and management science. Engineers will also find this book useful.




Cycle Representations of Markov Processes


Book Description

This book provides new insight into Markovian dependence via the cycle decompositions. It presents a systematic account of a class of stochastic processes known as cycle (or circuit) processes - so-called because they may be defined by directed cycles. An important application of this approach is the insight it provides to electrical networks and the duality principle of networks. This expanded second edition adds new advances, which reveal wide-ranging interpretations of cycle representations such as homologic decompositions, orthogonality equations, Fourier series, semigroup equations, and disintegration of measures. The text includes chapter summaries as well as a number of detailed illustrations.




Stochastic Modeling


Book Description

Three coherent parts form the material covered in this text, portions of which have not been widely covered in traditional textbooks. In this coverage the reader is quickly introduced to several different topics enriched with 175 exercises which focus on real-world problems. Exercises range from the classics of probability theory to more exotic research-oriented problems based on numerical simulations. Intended for graduate students in mathematics and applied sciences, the text provides the tools and training needed to write and use programs for research purposes. The first part of the text begins with a brief review of measure theory and revisits the main concepts of probability theory, from random variables to the standard limit theorems. The second part covers traditional material on stochastic processes, including martingales, discrete-time Markov chains, Poisson processes, and continuous-time Markov chains. The theory developed is illustrated by a variety of examples surrounding applications such as the gambler’s ruin chain, branching processes, symmetric random walks, and queueing systems. The third, more research-oriented part of the text, discusses special stochastic processes of interest in physics, biology, and sociology. Additional emphasis is placed on minimal models that have been used historically to develop new mathematical techniques in the field of stochastic processes: the logistic growth process, the Wright –Fisher model, Kingman’s coalescent, percolation models, the contact process, and the voter model. Further treatment of the material explains how these special processes are connected to each other from a modeling perspective as well as their simulation capabilities in C and MatlabTM.







Controlled Markov Processes and Viscosity Solutions


Book Description

This book is an introduction to optimal stochastic control for continuous time Markov processes and the theory of viscosity solutions. It covers dynamic programming for deterministic optimal control problems, as well as to the corresponding theory of viscosity solutions. New chapters in this second edition introduce the role of stochastic optimal control in portfolio optimization and in pricing derivatives in incomplete markets and two-controller, zero-sum differential games.




Stochastic Modeling


Book Description

Coherent introduction to techniques also offers a guide to the mathematical, numerical, and simulation tools of systems analysis. Includes formulation of models, analysis, and interpretation of results. 1995 edition.




Markov Processes for Stochastic Modeling


Book Description

This book presents an algebraic development of the theory of countable state space Markov chains with discrete- and continuous-time parameters. A Markov chain is a stochastic process characterized by the Markov prop erty that the distribution of future depends only on the current state, not on the whole history. Despite its simple form of dependency, the Markov property has enabled us to develop a rich system of concepts and theorems and to derive many results that are useful in applications. In fact, the areas that can be modeled, with varying degrees of success, by Markov chains are vast and are still expanding. The aim of this book is a discussion of the time-dependent behavior, called the transient behavior, of Markov chains. From the practical point of view, when modeling a stochastic system by a Markov chain, there are many instances in which time-limiting results such as stationary distributions have no meaning. Or, even when the stationary distribution is of some importance, it is often dangerous to use the stationary result alone without knowing the transient behavior of the Markov chain. Not many books have paid much attention to this topic, despite its obvious importance.




Markov Models & Optimization


Book Description

This book presents a radically new approach to problems of evaluating and optimizing the performance of continuous-time stochastic systems. This approach is based on the use of a family of Markov processes called Piecewise-Deterministic Processes (PDPs) as a general class of stochastic system models. A PDP is a Markov process that follows deterministic trajectories between random jumps, the latter occurring either spontaneously, in a Poisson-like fashion, or when the process hits the boundary of its state space. This formulation includes an enormous variety of applied problems in engineering, operations research, management science and economics as special cases; examples include queueing systems, stochastic scheduling, inventory control, resource allocation problems, optimal planning of production or exploitation of renewable or non-renewable resources, insurance analysis, fault detection in process systems, and tracking of maneuvering targets, among many others. The first part of the book shows how these applications lead to the PDP as a system model, and the main properties of PDPs are derived. There is particular emphasis on the so-called extended generator of the process, which gives a general method for calculating expectations and distributions of system performance functions. The second half of the book is devoted to control theory for PDPs, with a view to controlling PDP models for optimal performance: characterizations are obtained of optimal strategies both for continuously-acting controllers and for control by intervention (impulse control). Throughout the book, modern methods of stochastic analysis are used, but all the necessary theory is developed from scratch and presented in a self-contained way. The book will be useful to engineers and scientists in the application areas as well as to mathematicians interested in applications of stochastic analysis.




Constrained Markov Decision Processes


Book Description

This book provides a unified approach for the study of constrained Markov decision processes with a finite state space and unbounded costs. Unlike the single controller case considered in many other books, the author considers a single controller with several objectives, such as minimizing delays and loss, probabilities, and maximization of throughputs. It is desirable to design a controller that minimizes one cost objective, subject to inequality constraints on other cost objectives. This framework describes dynamic decision problems arising frequently in many engineering fields. A thorough overview of these applications is presented in the introduction. The book is then divided into three sections that build upon each other.