Practical Methods for Optimal Control Using Nonlinear Programming, Third Edition


Book Description

How do you fly an airplane from one point to another as fast as possible? What is the best way to administer a vaccine to fight the harmful effects of disease? What is the most efficient way to produce a chemical substance? This book presents practical methods for solving real optimal control problems such as these. Practical Methods for Optimal Control Using Nonlinear Programming, Third Edition focuses on the direct transcription method for optimal control. It features a summary of relevant material in constrained optimization, including nonlinear programming; discretization techniques appropriate for ordinary differential equations and differential-algebraic equations; and several examples and descriptions of computational algorithm formulations that implement this discretize-then-optimize strategy. The third edition has been thoroughly updated and includes new material on implicit Runge–Kutta discretization techniques, new chapters on partial differential equations and delay equations, and more than 70 test problems and open source FORTRAN code for all of the problems. This book will be valuable for academic and industrial research and development in optimal control theory and applications. It is appropriate as a primary or supplementary text for advanced undergraduate and graduate students.







A Variational Approach to Optimal Control of ODEs


Book Description

This self-contained book presents in a unified, systematic way the basic principles of optimal control governed by ODEs. Using a variational perspective, the author incorporates important restrictions like constraints for control and state, as well as the state system itself, into the equivalent variational reformulation of the problem. The fundamental issues of existence of optimal solutions, optimality conditions, and numerical approximation are then examined from this variational viewpoint. Inside, readers will find a unified approach to all the basic issues of optimal control, academic and real-world examples testing the book’s variational approach, and a rigorous treatment stressing ideas and arguments rather than the underlying mathematical formalism. A Variational Approach to Optimal Control of ODEs is mainly for applied analysts, applied mathematicians, and control engineers, but will also be helpful to other scientists and engineers who want to understand the basic principles of optimal control governed by ODEs. It requires no prerequisites in variational problems or expertise in numerical approximation. It can be used for a first course in optimal control.




Business Dynamics Models


Book Description

This book introduces optimal control methods, formulated as optimization problems, applied to business dynamics problems. Business dynamics refers to a combination of business management and financial objectives embedded in a dynamical system model. The model is subject to a control that optimizes a performance index and takes both management and financial aspects into account. Business Dynamics Models: Optimization-Based One Step Ahead Optimal Control includes solutions that provide a rationale for the use of optimal control and guidelines for further investigation into more complex models, as well as formulations that can also be used in a so-called flight simulator mode to investigate different complex scenarios. The text offers a modern programming environment (Jupyter notebooks in JuMP/Julia) for modeling, simulation, and optimization, and Julia code and notebooks are provided on a website for readers to experiment with their own examples. This book is intended for students majoring in applied mathematics, business, and engineering. The authors use a formulation-algorithm-example approach, rather than the classical definition-theorem-proof, making the material understandable to senior undergraduates and beginning graduates.




Extremum Seeking Through Delays and PDEs


Book Description

Extremum Seeking through Delays and PDEs, the first book on the topic, expands the scope of applicability of the extremum seeking method, from static and finite-dimensional systems to infinite-dimensional systems. Readers will find numerous algorithms for model-free real-time optimization are developed and their convergence guaranteed, extensions from single-player optimization to noncooperative games, under delays and PDEs, are provided, the delays and PDEs are compensated in the control designs using the PDE backstepping approach, and stability is ensured using infinite-dimensional versions of averaging theory, and accessible and powerful tools for analysis. This book is intended for control engineers in all disciplines (electrical, mechanical, aerospace, chemical), mathematicians, physicists, biologists, and economists. It is appropriate for graduate students, researchers, and industrial users.




Transfinite Interpolation and Eulerian/Lagrangian Dynamics


Book Description

This book introduces transfinite interpolation as a generalization of interpolation of data prescribed at a finite number of points to data prescribed on a geometrically structured set, such as a piece of curve, surface, or submanifold. The time-independent theory is readily extended to a moving/deforming data set whose dynamics is specified in a Eulerian or Lagrangian framework. The resulting innovative tools cover a very broad spectrum of applications in fluid mechanics, geometric optimization, and imaging. The authors chose to focus on the dynamical mesh updating in fluid mechanics and the construction of velocity fields from the boundary expression of the shape derivative. Transfinite Interpolations and Eulerian/Lagrangian Dynamics is a self-contained graduate-level text that integrates theory, applications, numerical approximations, and computational techniques. It applies transfinite interpolation methods to finite element mesh adaptation and ALE fluid-structure interaction. Specialists in applied mathematics, physics, mechanics, computational sciences, imaging sciences, and engineering will find this book of interest.




Observability


Book Description

This book is about nonlinear observability. It provides a modern theory of observability based on a new paradigm borrowed from theoretical physics and the mathematical foundation of that paradigm. In the case of observability, this framework takes into account the group of invariance that is inherent to the concept of observability, allowing the reader to reach an intuitive derivation of significant results in the literature of control theory. The book provides a complete theory of observability and, consequently, the analytical solution of some open problems in control theory. Notably, it presents the first general analytic solution of the nonlinear unknown input observability (nonlinear UIO), a very complex open problem studied in the 1960s. Based on this solution, the book provides examples with important applications for neuroscience, including a deep study of the integration of multiple sensory cues from the visual and vestibular systems for self-motion perception. Observability: A New Theory Based on the Group of Invariance is the only book focused solely on observability. It provides readers with many applications, mostly in robotics and autonomous navigation, as well as complex examples in the framework of vision-aided inertial navigation for aerial vehicles. For these applications, it also includes all the derivations needed to separate the observable part of the system from the unobservable, an analysis with practical importance for obtaining the basic equations for implementing any estimation scheme or for achieving a closed-form solution to the problem. This book is intended for researchers in robotics and automation, both in academia and in industry. Researchers in other engineering disciplines, such as information theory and mechanics, will also find the book useful.




Lectures on Stochastic Programming


Book Description

Optimization problems involving stochastic models occur in almost all areas of science and engineering, such as telecommunications, medicine, and finance. Their existence compels a need for rigorous ways of formulating, analyzing, and solving such problems. This book focuses on optimization problems involving uncertain parameters and covers the theoretical foundations and recent advances in areas where stochastic models are available. Readers will find coverage of the basic concepts of modeling these problems, including recourse actions and the nonanticipativity principle. The book also includes the theory of two-stage and multistage stochastic programming problems; the current state of the theory on chance (probabilistic) constraints, including the structure of the problems, optimality theory, and duality; and statistical inference in and risk-averse approaches to stochastic programming.




Numerical Control: Part B


Book Description

Numerical Control: Part B, Volume 24 in the Handbook of Numerical Analysis series, highlights new advances in the field, with this new volume presenting interesting chapters written by an international board of authors. Chapters in this volume include Control problems in the coefficients and the domain for linear elliptic equations, Computational approaches for extremal geometric eigenvalue problems, Non-overlapping domain decomposition in space and time for PDE-constrained optimal control problems on networks, Feedback Control of Time-dependent Nonlinear PDEs with Applications in Fluid Dynamics, Stabilization of the Navier-Stokes equations - Theoretical and numerical aspects, Reconstruction algorithms based on Carleman estimates, and more. Other sections cover Discrete time formulations as time discretization strategies in data assimilation, Back and forth iterations/Time reversal methods, Unbalanced Optimal Transport: from Theory to Numerics, An ADMM Approach to the Exact and Approximate Controllability of Parabolic Equations, Nonlocal balance laws -- an overview over recent results, Numerics and control of conservation laws, Numerical approaches for simulation and control of superconducting quantum circuits, and much more. - Provides the authority and expertise of leading contributors from an international board of authors - Presents the latest release in the Handbook of Numerical Analysis series - Updated release includes the latest information on Numerical Control




Computational Optimal Control


Book Description

Computational Optimal Control: Tools and Practice provides a detailed guide to informed use of computational optimal control in advanced engineering practice, addressing the need for a better understanding of the practical application of optimal control using computational techniques. Throughout the text the authors employ an advanced aeronautical case study to provide a practical, real-life setting for optimal control theory. This case study focuses on an advanced, real-world problem known as the “terminal bunt manoeuvre” or special trajectory shaping of a cruise missile. Representing the many problems involved in flight dynamics, practical control and flight path constraints, this case study offers an excellent illustration of advanced engineering practice using optimal solutions. The book describes in practical detail the real and tested optimal control software, examining the advantages and limitations of the technology. Featuring tutorial insights into computational optimal formulations and an advanced case-study approach to the topic, Computational Optimal Control: Tools and Practice provides an essential handbook for practising engineers and academics interested in practical optimal solutions in engineering. Focuses on an advanced, real-world aeronautical case study examining optimisation of the bunt manoeuvre Covers DIRCOL, NUDOCCCS, PROMIS and SOCS (under the GESOP environment), and BNDSCO Explains how to configure and optimize software to solve complex real-world computational optimal control problems Presents a tutorial three-stage hybrid approach to solving optimal control problem formulations