Algorithms for Smooth Nonconvex Optimization with Worst-case Guarantees


Book Description

The nature of global convergence guarantees for nonconvex optimization algorithms has changed significantly in recent years. New results characterize the maximum computational cost required for algorithms to satisfy approximate optimality conditions, instead of focusing on the limiting behavior of the iterates. In many contexts, such as those arising from machine learning, convergence to approximate second order points is desired. Algorithms designed for these problems must avoid saddle points efficiently to achieve optimal worst-case guarantees. In this dissertation, we develop and analyze a number of nonconvex optimization algorithms. First, we focus on accelerated gradient algorithms and provide results related to the avoidance of "strict saddle points''. In addition, the rate of divergence these accelerated gradient algorithms exhibit when in a neighborhood of strict saddle points is proven. Subsequently, we propose three new algorithms for smooth, nonconvex optimization with worst-case complexity guarantees. The first algorithm is developed for unconstrained optimization and is based on the classical Newton Conjugate Gradient method. This approach is then extended to bound constrained optimization by modifying the primal-log barrier method. Finally, we present a method for a special class of ``strict saddle functions'' which does not require knowledge of the parameters defining the optimization landscape. These algorithms converge to approximate second-order points in the best known computational complexity for their respective problem classes.




Structure and Complexity in Non-convex and Non-smooth Optimization


Book Description

Complexity theory drives much of modern optimization, allowing a fair comparison between competing numerical methods. The subject broadly seeks to both develop efficient algorithms and establish limitations on efficiencies of any algorithm for the problem class. Classical complexity theory based on oracle models targets problems that are both smooth and convex. Without smoothness, methods rely on exploiting the structure of the target function to improve on the worst-case complexity of non-smooth convex optimization. This thesis explores complexity of first-order methods for structured non-smooth and non-convex problems. A central example is the minimization of a composition of a convex function with a smooth map - the so-called convex-composite problem class. Nonlinear least squares formulations in engineering and nonlinear model fitting in statistics fall within this framework. The thesis develops new algorithms for the composite problem class, along with inertial variants that are adaptive to convexity. Acceleration is a widely used term in contemporary optimization. The term is often used to describe methods with efficiency guarantees matching the best possible complexity estimates for a given problem class. This thesis develops methods that interpolate between convex and non-convex settings. In particular, we focus on minimizing large finite sum problems, popular for modeling empirical risk in statistical applications, when the user is unaware of the convexity of the objective function. The scheme we describe has convergence guarantees that adapt to the underlying convexity of the objective function. First-order algorithms for non-smooth problems depend on having access to generalized derivatives of the objective function. We conclude the thesis with a fresh look at variational properties of spectral function. These are the functions on the space of symmetric matrices that depend on the matrix only through its eigenvalues. In particular, our analysis dramatically simplifies currently available derivations of differential formulas of such functions.




Evaluation Complexity of Algorithms for Nonconvex Optimization


Book Description

A popular way to assess the “effort” needed to solve a problem is to count how many evaluations of the problem functions (and their derivatives) are required. In many cases, this is often the dominating computational cost. Given an optimization problem satisfying reasonable assumptions—and given access to problem-function values and derivatives of various degrees—how many evaluations might be required to approximately solve the problem? Evaluation Complexity of Algorithms for Nonconvex Optimization: Theory, Computation, and Perspectives addresses this question for nonconvex optimization problems, those that may have local minimizers and appear most often in practice. This is the first book on complexity to cover topics such as composite and constrained optimization, derivative-free optimization, subproblem solution, and optimal (lower and sharpness) bounds for nonconvex problems. It is also the first to address the disadvantages of traditional optimality measures and propose useful surrogates leading to algorithms that compute approximate high-order critical points, and to compare traditional and new methods, highlighting the advantages of the latter from a complexity point of view. This is the go-to book for those interested in solving nonconvex optimization problems. It is suitable for advanced undergraduate and graduate students in courses on advanced numerical analysis, data science, numerical optimization, and approximation theory.







Trust Region Methods


Book Description

Mathematics of Computing -- General.







Lectures on Convex Optimization


Book Description

This book provides a comprehensive, modern introduction to convex optimization, a field that is becoming increasingly important in applied mathematics, economics and finance, engineering, and computer science, notably in data science and machine learning. Written by a leading expert in the field, this book includes recent advances in the algorithmic theory of convex optimization, naturally complementing the existing literature. It contains a unified and rigorous presentation of the acceleration techniques for minimization schemes of first- and second-order. It provides readers with a full treatment of the smoothing technique, which has tremendously extended the abilities of gradient-type methods. Several powerful approaches in structural optimization, including optimization in relative scale and polynomial-time interior-point methods, are also discussed in detail. Researchers in theoretical optimization as well as professionals working on optimization problems will find this book very useful. It presents many successful examples of how to develop very fast specialized minimization algorithms. Based on the author’s lectures, it can naturally serve as the basis for introductory and advanced courses in convex optimization for students in engineering, economics, computer science and mathematics.




High-Dimensional Data Analysis with Low-Dimensional Models


Book Description

Connects fundamental mathematical theory with real-world problems, through efficient and scalable optimization algorithms.




Multi-agent Optimization


Book Description

This book contains three well-written research tutorials that inform the graduate reader about the forefront of current research in multi-agent optimization. These tutorials cover topics that have not yet found their way in standard books and offer the reader the unique opportunity to be guided by major researchers in the respective fields. Multi-agent optimization, lying at the intersection of classical optimization, game theory, and variational inequality theory, is at the forefront of modern optimization and has recently undergone a dramatic development. It seems timely to provide an overview that describes in detail ongoing research and important trends. This book concentrates on Distributed Optimization over Networks; Differential Variational Inequalities; and Advanced Decomposition Algorithms for Multi-agent Systems. This book will appeal to both mathematicians and mathematically oriented engineers and will be the source of inspiration for PhD students and researchers.




Handbook of Intelligent Computing and Optimization for Sustainable Development


Book Description

HANDBOOK OF INTELLIGENT COMPUTING AND OPTIMIZATION FOR SUSTAINABLE DEVELOPMENT This book provides a comprehensive overview of the latest breakthroughs and recent progress in sustainable intelligent computing technologies, applications, and optimization techniques across various industries. Optimization has received enormous attention along with the rapidly increasing use of communication technology and the development of user-friendly software and artificial intelligence. In almost all human activities, there is a desire to deliver the highest possible results with the least amount of effort. Moreover, optimization is a very well-known area with a vast number of applications, from route finding problems to medical treatment, construction, finance, accounting, engineering, and maintenance schedules in plants. As far as optimization of real-world problems is concerned, understanding the nature of the problem and grouping it in a proper class may help the designer employ proper techniques which can solve the problem efficiently. Many intelligent optimization techniques can find optimal solutions without the use of objective function and are less prone to local conditions. The 41 chapters comprising the Handbook of Intelligent Computing and Optimization for Sustainable Development by subject specialists, represent diverse disciplines such as mathematics and computer science, electrical and electronics engineering, neuroscience and cognitive sciences, medicine, and social sciences, and provide the reader with an integrated understanding of the importance that intelligent computing has in the sustainable development of current societies. It discusses the emerging research exploring the theoretical and practical aspects of successfully implementing new and innovative intelligent techniques in a variety of sectors, including IoT, manufacturing, optimization, and healthcare. Audience It is a pivotal reference source for IT specialists, industry professionals, managers, executives, researchers, scientists, and engineers seeking current research in emerging perspectives in the field of artificial intelligence in the areas of Internet of Things, renewable energy, optimization, and smart cities.