Deep Neural Networks


Book Description

Toward Deep Neural Networks: WASD Neuronet Models, Algorithms, and Applications introduces the outlook and extension toward deep neural networks, with a focus on the weights-and-structure determination (WASD) algorithm. Based on the authors’ 20 years of research experience on neuronets, the book explores the models, algorithms, and applications of the WASD neuronet, and allows reader to extend the techniques in the book to solve scientific and engineering problems. The book will be of interest to engineers, senior undergraduates, postgraduates, and researchers in the fields of neuronets, computer mathematics, computer science, artificial intelligence, numerical algorithms, optimization, simulation and modeling, deep learning, and data mining. Features Focuses on neuronet models, algorithms, and applications Designs, constructs, develops, analyzes, simulates and compares various WASD neuronet models, such as single-input WASD neuronet models, two-input WASD neuronet models, three-input WASD neuronet models, and general multi-input WASD neuronet models for function data approximations Includes real-world applications, such as population prediction Provides complete mathematical foundations, such as Weierstrass approximation, Bernstein polynomial approximation, Taylor polynomial approximation, and multivariate function approximation, exploring the close integration of mathematics (i.e., function approximation theories) and computers (e.g., computer algorithms) Utilizes the authors' 20 years of research on neuronets




Toward Deep Neural Networks


Book Description

Toward Deep Neural Networks: WASD Neuronet Models, Algorithms, and Applications introduces the outlook and extension toward deep neural networks, with a focus on the weights-and-structure determination (WASD) algorithm. Based on the authors’ 20 years of research experience on neuronets, the book explores the models, algorithms, and applications of the WASD neuronet, and allows reader to extend the techniques in the book to solve scientific and engineering problems. The book will be of interest to engineers, senior undergraduates, postgraduates, and researchers in the fields of neuronets, computer mathematics, computer science, artificial intelligence, numerical algorithms, optimization, simulation and modeling, deep learning, and data mining. Features Focuses on neuronet models, algorithms, and applications Designs, constructs, develops, analyzes, simulates and compares various WASD neuronet models, such as single-input WASD neuronet models, two-input WASD neuronet models, three-input WASD neuronet models, and general multi-input WASD neuronet models for function data approximations Includes real-world applications, such as population prediction Provides complete mathematical foundations, such as Weierstrass approximation, Bernstein polynomial approximation, Taylor polynomial approximation, and multivariate function approximation, exploring the close integration of mathematics (i.e., function approximation theories) and computers (e.g., computer algorithms) Utilizes the authors' 20 years of research on neuronets




The Principles of Deep Learning Theory


Book Description

This volume develops an effective theory approach to understanding deep neural networks of practical relevance.




Automated Machine Learning


Book Description

This open access book presents the first comprehensive overview of general methods in Automated Machine Learning (AutoML), collects descriptions of existing systems based on these methods, and discusses the first series of international challenges of AutoML systems. The recent success of commercial ML applications and the rapid growth of the field has created a high demand for off-the-shelf ML methods that can be used easily and without expert knowledge. However, many of the recent machine learning successes crucially rely on human experts, who manually select appropriate ML architectures (deep learning architectures or more traditional ML workflows) and their hyperparameters. To overcome this problem, the field of AutoML targets a progressive automation of machine learning, based on principles from optimization and machine learning itself. This book serves as a point of entry into this quickly-developing field for researchers and advanced students alike, as well as providing a reference for practitioners aiming to use AutoML in their work.




Efficient Processing of Deep Neural Networks


Book Description

This book provides a structured treatment of the key principles and techniques for enabling efficient processing of deep neural networks (DNNs). DNNs are currently widely used for many artificial intelligence (AI) applications, including computer vision, speech recognition, and robotics. While DNNs deliver state-of-the-art accuracy on many AI tasks, it comes at the cost of high computational complexity. Therefore, techniques that enable efficient processing of deep neural networks to improve key metrics—such as energy-efficiency, throughput, and latency—without sacrificing accuracy or increasing hardware costs are critical to enabling the wide deployment of DNNs in AI systems. The book includes background on DNN processing; a description and taxonomy of hardware architectural approaches for designing DNN accelerators; key metrics for evaluating and comparing different designs; features of DNN processing that are amenable to hardware/algorithm co-design to improve energy efficiency and throughput; and opportunities for applying new technologies. Readers will find a structured introduction to the field as well as formalization and organization of key concepts from contemporary work that provide insights that may spark new ideas.




Strengthening Deep Neural Networks


Book Description

As deep neural networks (DNNs) become increasingly common in real-world applications, the potential to deliberately "fool" them with data that wouldn’t trick a human presents a new attack vector. This practical book examines real-world scenarios where DNNs—the algorithms intrinsic to much of AI—are used daily to process image, audio, and video data. Author Katy Warr considers attack motivations, the risks posed by this adversarial input, and methods for increasing AI robustness to these attacks. If you’re a data scientist developing DNN algorithms, a security architect interested in how to make AI systems more resilient to attack, or someone fascinated by the differences between artificial and biological perception, this book is for you. Delve into DNNs and discover how they could be tricked by adversarial input Investigate methods used to generate adversarial input capable of fooling DNNs Explore real-world scenarios and model the adversarial threat Evaluate neural network robustness; learn methods to increase resilience of AI systems to adversarial data Examine some ways in which AI might become better at mimicking human perception in years to come




Introduction to Neural Network Verification


Book Description

Over the past decade, a number of hardware and software advances have conspired to thrust deep learning and neural networks to the forefront of computing. Deep learning has created a qualitative shift in our conception of what software is and what it can do: Every day we're seeing new applications of deep learning, from healthcare to art, and it feels like we're only scratching the surface of a universe of new possibilities. This book offers the first introduction of foundational ideas from automated verification as applied to deep neural networks and deep learning. It is divided into three parts: Part 1 defines neural networks as data-flow graphs of operators over real-valued inputs. Part 2 discusses constraint-based techniques for verification. Part 3 discusses abstraction-based techniques for verification. The book is a self-contained treatment of a topic that sits at the intersection of machine learning and formal verification. It can serve as an introduction to the field for first-year graduate students or senior undergraduates, even if they have not been exposed to deep learning or verification.




Deep Neural Networks in a Mathematical Framework


Book Description

This SpringerBrief describes how to build a rigorous end-to-end mathematical framework for deep neural networks. The authors provide tools to represent and describe neural networks, casting previous results in the field in a more natural light. In particular, the authors derive gradient descent algorithms in a unified way for several neural network structures, including multilayer perceptrons, convolutional neural networks, deep autoencoders and recurrent neural networks. Furthermore, the authors developed framework is both more concise and mathematically intuitive than previous representations of neural networks. This SpringerBrief is one step towards unlocking the black box of Deep Learning. The authors believe that this framework will help catalyze further discoveries regarding the mathematical properties of neural networks.This SpringerBrief is accessible not only to researchers, professionals and students working and studying in the field of deep learning, but also to those outside of the neutral network community.




Learning Deep Architectures for AI


Book Description

Theoretical results suggest that in order to learn the kind of complicated functions that can represent high-level abstractions (e.g. in vision, language, and other AI-level tasks), one may need deep architectures. Deep architectures are composed of multiple levels of non-linear operations, such as in neural nets with many hidden layers or in complicated propositional formulae re-using many sub-formulae. Searching the parameter space of deep architectures is a difficult task, but learning algorithms such as those for Deep Belief Networks have recently been proposed to tackle this problem with notable success, beating the state-of-the-art in certain areas. This paper discusses the motivations and principles regarding learning algorithms for deep architectures, in particular those exploiting as building blocks unsupervised learning of single-layer models such as Restricted Boltzmann Machines, used to construct deeper models such as Deep Belief Networks.




Generative Deep Learning


Book Description

Generative modeling is one of the hottest topics in AI. It’s now possible to teach a machine to excel at human endeavors such as painting, writing, and composing music. With this practical book, machine-learning engineers and data scientists will discover how to re-create some of the most impressive examples of generative deep learning models, such as variational autoencoders,generative adversarial networks (GANs), encoder-decoder models and world models. Author David Foster demonstrates the inner workings of each technique, starting with the basics of deep learning before advancing to some of the most cutting-edge algorithms in the field. Through tips and tricks, you’ll understand how to make your models learn more efficiently and become more creative. Discover how variational autoencoders can change facial expressions in photos Build practical GAN examples from scratch, including CycleGAN for style transfer and MuseGAN for music generation Create recurrent generative models for text generation and learn how to improve the models using attention Understand how generative models can help agents to accomplish tasks within a reinforcement learning setting Explore the architecture of the Transformer (BERT, GPT-2) and image generation models such as ProGAN and StyleGAN