Linear Systems Control


Book Description

Modern control theory and in particular state space or state variable methods can be adapted to the description of many different systems because it depends strongly on physical modeling and physical intuition. The laws of physics are in the form of differential equations and for this reason, this book concentrates on system descriptions in this form. This means coupled systems of linear or nonlinear differential equations. The physical approach is emphasized in this book because it is most natural for complex systems. It also makes what would ordinarily be a difficult mathematical subject into one which can straightforwardly be understood intuitively and which deals with concepts which engineering and science students are already familiar. In this way it is easy to immediately apply the theory to the understanding and control of ordinary systems. Application engineers, working in industry, will also find this book interesting and useful for this reason. In line with the approach set forth above, the book first deals with the modeling of systems in state space form. Both transfer function and differential equation modeling methods are treated with many examples. Linearization is treated and explained first for very simple nonlinear systems and then more complex systems. Because computer control is so fundamental to modern applications, discrete time modeling of systems as difference equations is introduced immediately after the more intuitive differential equation models. The conversion of differential equation models to difference equations is also discussed at length, including transfer function formulations. A vital problem in modern control is how to treat noise in control systems. Nevertheless this question is rarely treated in many control system textbooks because it is considered to be too mathematical and too difficult in a second course on controls. In this textbook a simple physical approach is made to the description of noise and stochastic disturbances which is easy to understand and apply to common systems. This requires only a few fundamental statistical concepts which are given in a simple introduction which lead naturally to the fundamental noise propagation equation for dynamic systems, the Lyapunov equation. This equation is given and exemplified both in its continuous and discrete time versions. With the Lyapunov equation available to describe state noise propagation, it is a very small step to add the effect of measurements and measurement noise. This gives immediately the Riccati equation for optimal state estimators or Kalman filters. These important observers are derived and illustrated using simulations in terms which make them easy to understand and easy to apply to real systems. The use of LQR regulators with Kalman filters give LQG (Linear Quadratic Gaussian) regulators which are introduced at the end of the book. Another important subject which is introduced is the use of Kalman filters as parameter estimations for unknown parameters. The textbook is divided into 7 chapters, 5 appendices, a table of contents, a table of examples, extensive index and extensive list of references. Each chapter is provided with a summary of the main points covered and a set of problems relevant to the material in that chapter. Moreover each of the more advanced chapters (3 - 7) are provided with notes describing the history of the mathematical and technical problems which lead to the control theory presented in that chapter. Continuous time methods are the main focus in the book because these provide the most direct connection to physics. This physical foundation allows a logical presentation and gives a good intuitive feel for control system construction. Nevertheless strong attention is also given to discrete time systems. Very few proofs are included in the book but most of the important results are derived. This method of presentation makes the text very readable and gives a good foundation for reading more rigorous texts. A complete set of solutions is available for all of the problems in the text. In addition a set of longer exercises is available for use as Matlab/Simulink ‘laboratory exercises’ in connection with lectures. There is material of this kind for 12 such exercises and each exercise requires about 3 hours for its solution. Full written solutions of all these exercises are available.




Progress on Difference Equations and Discrete Dynamical Systems


Book Description

This book comprises selected papers of the 25th International Conference on Difference Equations and Applications, ICDEA 2019, held at UCL, London, UK, in June 2019. The volume details the latest research on difference equations and discrete dynamical systems, and their application to areas such as biology, economics, and the social sciences. Some chapters have a tutorial style and cover the history and more recent developments for a particular topic, such as chaos, bifurcation theory, monotone dynamics, and global stability. Other chapters cover the latest personal research contributions of the author(s) in their particular area of expertise and range from the more technical articles on abstract systems to those that discuss the application of difference equations to real-world problems. The book is of interest to both Ph.D. students and researchers alike who wish to keep abreast of the latest developments in difference equations and discrete dynamical systems.




Mathematical Methods in Robust Control of Discrete-Time Linear Stochastic Systems


Book Description

In this monograph the authors develop a theory for the robust control of discrete-time stochastic systems, subjected to both independent random perturbations and to Markov chains. Such systems are widely used to provide mathematical models for real processes in fields such as aerospace engineering, communications, manufacturing, finance and economy. The theory is a continuation of the authors’ work presented in their previous book entitled "Mathematical Methods in Robust Control of Linear Stochastic Systems" published by Springer in 2006. Key features: - Provides a common unifying framework for discrete-time stochastic systems corrupted with both independent random perturbations and with Markovian jumps which are usually treated separately in the control literature; - Covers preliminary material on probability theory, independent random variables, conditional expectation and Markov chains; - Proposes new numerical algorithms to solve coupled matrix algebraic Riccati equations; - Leads the reader in a natural way to the original results through a systematic presentation; - Presents new theoretical results with detailed numerical examples. The monograph is geared to researchers and graduate students in advanced control engineering, applied mathematics, mathematical systems theory and finance. It is also accessible to undergraduate students with a fundamental knowledge in the theory of stochastic systems.




Stability of Dynamical Systems


Book Description

In the analysis and synthesis of contemporary systems, engineers and scientists are frequently confronted with increasingly complex models that may simultaneously include components whose states evolve along continuous time and discrete instants; components whose descriptions may exhibit nonlinearities, time lags, transportation delays, hysteresis effects, and uncertainties in parameters; and components that cannot be described by various classical equations, as in the case of discrete-event systems, logic commands, and Petri nets. The qualitative analysis of such systems requires results for finite-dimensional and infinite-dimensional systems; continuous-time and discrete-time systems; continuous continuous-time and discontinuous continuous-time systems; and hybrid systems involving a mixture of continuous and discrete dynamics. Filling a gap in the literature, this textbook presents the first comprehensive stability analysis of all the major types of system models described above. Throughout the book, the applicability of the developed theory is demonstrated by means of many specific examples and applications to important classes of systems, including digital control systems, nonlinear regulator systems, pulse-width-modulated feedback control systems, artificial neural networks (with and without time delays), digital signal processing, a class of discrete-event systems (with applications to manufacturing and computer load balancing problems) and a multicore nuclear reactor model. The book covers the following four general topics: * Representation and modeling of dynamical systems of the types described above * Presentation of Lyapunov and Lagrange stability theory for dynamical systems defined on general metric spaces * Specialization of this stability theory to finite-dimensional dynamical systems * Specialization of this stability theory to infinite-dimensional dynamical systems Replete with exercises and requiring basic knowledge of linear algebra, analysis, and differential equations, the work may be used as a textbook for graduate courses in stability theory of dynamical systems. The book may also serve as a self-study reference for graduate students, researchers, and practitioners in applied mathematics, engineering, computer science, physics, chemistry, biology, and economics.




Modeling and Control of Logical Discrete Event Systems


Book Description

The field of discrete event systems has emerged to provide a formal treatment of many of the man-made systems such as manufacturing systems, communica tion networks. automated traffic systems, database management systems, and computer systems that are event-driven, highly complex, and not amenable to the classical treatments based on differential or difference equations. Discrete event systems is a growing field that utilizes many interesting mathematical models and techniques. In this book we focus on a high level treatment of discrete event systems. where the order of events. rather than their occurrence times, is the principal concern. Such treatment is needed to guarantee that the system under study meets desired logical goals. In this framework, dis crete event systems are modeled by formal languages or, equivalently, by state machines. The field of logical discrete event systems is an interdisciplinary field-it in cludes ideas from computer science, control theory, and operations research. Our goal is to bring together in one book the relevant techniques from these fields. This is the first book of this kind, and our hope is that it will be useful to professionals in the area of discrete event systems since most of the material presented has appeared previously only in journals. The book is also designed for a graduate level course on logical discrete event systems. It contains all the necessary background material in formal language theory and lattice the ory. The only prerequisite is some degree of "mathematical maturity".










Stability and Stable Oscillations in Discrete Time Systems


Book Description

The expertise of a professional mathmatician and a theoretical engineer provides a fresh perspective of stability and stable oscillations. The current state of affairs in stability theory, absolute stability of control systems, and stable oscillations of both periodic and almost periodic discrete systems is presented, including many applications in engineering such as stability of digital filters, digitally controlled thermal processes, neurodynamics, and chemical kinetics. This book will be an invaluable reference source for those whose work is in the area of discrete dynamical systems, difference equations, and control theory or applied areas that use discrete time models.




Impulsive Control in Continuous and Discrete-Continuous Systems


Book Description

Impulsive Control in Continuous and Discrete-Continuous Systems is an up-to-date introduction to the theory of impulsive control in nonlinear systems. This is a new branch of the Optimal Control Theory, which is tightly connected to the Theory of Hybrid Systems. The text introduces the reader to the interesting area of optimal control problems with discontinuous solutions, discussing the application of a new and effective method of discontinuous time-transformation. With a large number of examples, illustrations, and applied problems arising in the area of observation control, this book is excellent as a textbook or reference for a senior or graduate-level course on the subject, as well as a reference for researchers in related fields.




Stochastic Differential Equations with Markovian Switching


Book Description

This textbook provides the first systematic presentation of the theory of stochastic differential equations with Markovian switching. It presents the basic principles at an introductory level but emphasizes current advanced level research trends. The material takes into account all the features of Ito equations, Markovian switching, interval systems and time-lag. The theory developed is applicable in different and complicated situations in many branches of science and industry.