From Shortest Paths to Reinforcement Learning


Book Description

Dynamic programming (DP) has a relevant history as a powerful and flexible optimization principle, but has a bad reputation as a computationally impractical tool. This book fills a gap between the statement of DP principles and their actual software implementation. Using MATLAB throughout, this tutorial gently gets the reader acquainted with DP and its potential applications, offering the possibility of actual experimentation and hands-on experience. The book assumes basic familiarity with probability and optimization, and is suitable to both practitioners and graduate students in engineering, applied mathematics, management, finance and economics.




Reinforcement Learning, second edition


Book Description

The significantly expanded and updated new edition of a widely used text on reinforcement learning, one of the most active research areas in artificial intelligence. Reinforcement learning, one of the most active research areas in artificial intelligence, is a computational approach to learning whereby an agent tries to maximize the total amount of reward it receives while interacting with a complex, uncertain environment. In Reinforcement Learning, Richard Sutton and Andrew Barto provide a clear and simple account of the field's key ideas and algorithms. This second edition has been significantly expanded and updated, presenting new topics and updating coverage of other topics. Like the first edition, this second edition focuses on core online learning algorithms, with the more mathematical material set off in shaded boxes. Part I covers as much of reinforcement learning as possible without going beyond the tabular case for which exact solutions can be found. Many algorithms presented in this part are new to the second edition, including UCB, Expected Sarsa, and Double Learning. Part II extends these ideas to function approximation, with new sections on such topics as artificial neural networks and the Fourier basis, and offers expanded treatment of off-policy learning and policy-gradient methods. Part III has new chapters on reinforcement learning's relationships to psychology and neuroscience, as well as an updated case-studies chapter including AlphaGo and AlphaGo Zero, Atari game playing, and IBM Watson's wagering strategy. The final chapter discusses the future societal impacts of reinforcement learning.




From Shortest Paths to Reinforcement Learning


Book Description

Dynamic programming (DP) has a relevant history as a powerful and flexible optimization principle, but has a bad reputation as a computationally impractical tool. This book fills a gap between the statement of DP principles and their actual software implementation. Using MATLAB throughout, this tutorial gently gets the reader acquainted with DP and its potential applications, offering the possibility of actual experimentation and hands-on experience. The book assumes basic familiarity with probability and optimization, and is suitable to both practitioners and graduate students in engineering, applied mathematics, management, finance and economics.




A Course in Reinforcement Learning


Book Description

These lecture notes were prepared for use in the 2023 ASU research-oriented course on Reinforcement Learning (RL) that I have offered in each of the last five years. Their purpose is to give an overview of the RL methodology, particularly as it relates to problems of optimal and suboptimal decision and control, as well as discrete optimization. There are two major methodological RL approaches: approximation in value space, where we approximate in some way the optimal value function, and approximation in policy space, whereby we construct a (generally suboptimal) policy by using optimization over a suitably restricted class of policies.The lecture notes focus primarily on approximation in value space, with limited coverage of approximation in policy space. However, they are structured so that they can be easily supplemented by an instructor who wishes to go into approximation in policy space in greater detail, using any of a number of available sources, including the author's 2019 RL book. While in these notes we deemphasize mathematical proofs, there is considerable related analysis, which supports our conclusions and can be found in the author's recent RL and DP books. These books also contain additional material on off-line training of neural networks, on the use of policy gradient methods for approximation in policy space, and on aggregation.




Recent Advances in Reinforcement Learning


Book Description

Recent Advances in Reinforcement Learning addresses current research in an exciting area that is gaining a great deal of popularity in the Artificial Intelligence and Neural Network communities. Reinforcement learning has become a primary paradigm of machine learning. It applies to problems in which an agent (such as a robot, a process controller, or an information-retrieval engine) has to learn how to behave given only information about the success of its current actions. This book is a collection of important papers that address topics including the theoretical foundations of dynamic programming approaches, the role of prior knowledge, and methods for improving performance of reinforcement-learning techniques. These papers build on previous work and will form an important resource for students and researchers in the area. Recent Advances in Reinforcement Learning is an edited volume of peer-reviewed original research comprising twelve invited contributions by leading researchers. This research work has also been published as a special issue of Machine Learning (Volume 22, Numbers 1, 2 and 3).




High-Performance Big Data Computing


Book Description

An in-depth overview of an emerging field that brings together high-performance computing, big data processing, and deep lLearning. Over the last decade, the exponential explosion of data known as big data has changed the way we understand and harness the power of data. The emerging field of high-performance big data computing, which brings together high-performance computing (HPC), big data processing, and deep learning, aims to meet the challenges posed by large-scale data processing. This book offers an in-depth overview of high-performance big data computing and the associated technical issues, approaches, and solutions. The book covers basic concepts and necessary background knowledge, including data processing frameworks, storage systems, and hardware capabilities; offers a detailed discussion of technical issues in accelerating big data computing in terms of computation, communication, memory and storage, codesign, workload characterization and benchmarking, and system deployment and management; and surveys benchmarks and workloads for evaluating big data middleware systems. It presents a detailed discussion of big data computing systems and applications with high-performance networking, computing, and storage technologies, including state-of-the-art designs for data processing and storage systems. Finally, the book considers some advanced research topics in high-performance big data computing, including designing high-performance deep learning over big data (DLoBD) stacks and HPC cloud technologies.




Wireless Communications and Networking for Unmanned Aerial Vehicles


Book Description

"The past few years witnessed a major revolution in the area of unmanned aerial vehicles (UAVs), commonly known as drones, due to significant technological advances across various drone-related fields ranging from embedded systems to autonomy, control, security, and communications. These unprecedented recent advances in UAV technology have made it possible to widely deploy drones across a plethora of application domains ranging from delivery of goods to surveillance, environmental monitoring, track control, remote sensing, and search and rescue. In fact, recent reports from the Federal Aviation Administration (FAA) anticipate that sales of UAVs may exceed 7 million in 2020 and many industries are currently investing in innovative drone-centric applications and research. To enable all such applications, it is imperative to address a plethora of research challenges pertaining to drone systems, ranging from navigation to autonomy, control, sensing, navigation, and communications. In particular, the deployment of UAVs in tomorrow's smart cities, is largely contingent upon equipping them with effective means for communications and networking. To this end, in this book, we provide a comprehensive treatment of the wireless communications and networking research challenges and opportunities associated with UAV technology. This treatment begins in this chapter which provides an introduction to UAV technology and an in-depth discussion on the wireless communication and networking challenges associated with the introduction of UAVs"--