Hardware-Aware Probabilistic Machine Learning Models


Book Description

This book proposes probabilistic machine learning models that represent the hardware properties of the device hosting them. These models can be used to evaluate the impact that a specific device configuration may have on resource consumption and performance of the machine learning task, with the overarching goal of balancing the two optimally. The book first motivates extreme-edge computing in the context of the Internet of Things (IoT) paradigm. Then, it briefly reviews the steps involved in the execution of a machine learning task and identifies the implications associated with implementing this type of workload in resource-constrained devices. The core of this book focuses on augmenting and exploiting the properties of Bayesian Networks and Probabilistic Circuits in order to endow them with hardware-awareness. The proposed models can encode the properties of various device sub-systems that are typically not considered by other resource-aware strategies, bringing about resource-saving opportunities that traditional approaches fail to uncover. The performance of the proposed models and strategies is empirically evaluated for several use cases. All of the considered examples show the potential of attaining significant resource-saving opportunities with minimal accuracy losses at application time. Overall, this book constitutes a novel approach to hardware-algorithm co-optimization that further bridges the fields of Machine Learning and Electrical Engineering.




IoT Streams for Data-Driven Predictive Maintenance and IoT, Edge, and Mobile for Embedded Machine Learning


Book Description

This book constitutes selected papers from the Second International Workshop on IoT Streams for Data-Driven Predictive Maintenance, IoT Streams 2020, and First International Workshop on IoT, Edge, and Mobile for Embedded Machine Learning, ITEM 2020, co-located with ECML/PKDD 2020 and held in September 2020. Due to the COVID-19 pandemic the workshops were held online. The 21 full papers and 3 short papers presented in this volume were thoroughly reviewed and selected from 35 submissions and are organized according to the workshops and their topics: IoT Streams 2020: Stream Learning; Feature Learning; ITEM 2020: Unsupervised Machine Learning; Hardware; Methods; Quantization.




Efficient Execution of Irregular Dataflow Graphs


Book Description

This book focuses on the acceleration of emerging irregular sparse workloads, posed by novel artificial intelligent (AI) models and sparse linear algebra. Specifically, the book outlines several co-optimized hardware-software solutions for a highly promising class of emerging sparse AI models called Probabilistic Circuit (PC) and a similar sparse matrix workload for triangular linear systems (SpTRSV). The authors describe optimizations for the entire stack, targeting applications, compilation, hardware architecture and silicon implementation, resulting in orders of magnitude higher performance and energy-efficiency compared to the existing state-of-the-art solutions. Thus, this book provides important building blocks for the upcoming generation of edge AI platforms.




Advances in Intelligent Data Analysis XVIII


Book Description

This open access book constitutes the proceedings of the 18th International Conference on Intelligent Data Analysis, IDA 2020, held in Konstanz, Germany, in April 2020. The 45 full papers presented in this volume were carefully reviewed and selected from 114 submissions. Advancing Intelligent Data Analysis requires novel, potentially game-changing ideas. IDA’s mission is to promote ideas over performance: a solid motivation can be as convincing as exhaustive empirical evaluation.




Fundamentals


Book Description

Machine learning is part of Artificial Intelligence since its beginning. Certainly, not learning would only allow the perfect being to show intelligent behavior. All others, be it humans or machines, need to learn in order to enhance their capabilities. In the eighties of the last century, learning from examples and modeling human learning strategies have been investigated in concert. The formal statistical basis of many learning methods has been put forward later on and is still an integral part of machine learning. Neural networks have always been in the toolbox of methods. Integrating all the pre-processing, exploitation of kernel functions, and transformation steps of a machine learning process into the architecture of a deep neural network increased the performance of this model type considerably. Modern machine learning is challenged on the one hand by the amount of data and on the other hand by the demand of real-time inference. This leads to an interest in computing architectures and modern processors. For a long time, the machine learning research could take the von-Neumann architecture for granted. All algorithms were designed for the classical CPU. Issues of implementation on a particular architecture have been ignored. This is no longer possible. The time for independently investigating machine learning and computational architecture is over. Computing architecture has experienced a similarly rampant development from mainframe or personal computers in the last century to now very large compute clusters on the one hand and ubiquitous computing of embedded systems in the Internet of Things on the other hand. Cyber-physical systems’ sensors produce a huge amount of streaming data which need to be stored and analyzed. Their actuators need to react in real-time. This clearly establishes a close connection with machine learning. Cyber-physical systems and systems in the Internet of Things consist of diverse components, heterogeneous both in hard- and software. Modern multi-core systems, graphic processors, memory technologies and hardware-software codesign offer opportunities for better implementations of machine learning models. Machine learning and embedded systems together now form a field of research which tackles leading edge problems in machine learning, algorithm engineering, and embedded systems. Machine learning today needs to make the resource demands of learning and inference meet the resource constraints of used computer architecture and platforms. A large variety of algorithms for the same learning method and, moreover, diverse implementations of an algorithm for particular computing architectures optimize learning with respect to resource efficiency while keeping some guarantees of accuracy. The trade-off between a decreased energy consumption and an increased error rate, to just give an example, needs to be theoretically shown for training a model and the model inference. Pruning and quantization are ways of reducing the resource requirements by either compressing or approximating the model. In addition to memory and energy consumption, timeliness is an important issue, since many embedded systems are integrated into large products that interact with the physical world. If the results are delivered too late, they may have become useless. As a result, real-time guarantees are needed for such systems. To efficiently utilize the available resources, e.g., processing power, memory, and accelerators, with respect to response time, energy consumption, and power dissipation, different scheduling algorithms and resource management strategies need to be developed. This book series addresses machine learning under resource constraints as well as the application of the described methods in various domains of science and engineering. Turning big data into smart data requires many steps of data analysis: methods for extracting and selecting features, filtering and cleaning the data, joining heterogeneous sources, aggregating the data, and learning predictions need to scale up. The algorithms are challenged on the one hand by high-throughput data, gigantic data sets like in astrophysics, on the other hand by high dimensions like in genetic data. Resource constraints are given by the relation between the demands for processing the data and the capacity of the computing machinery. The resources are runtime, memory, communication, and energy. Novel machine learning algorithms are optimized with regard to minimal resource consumption. Moreover, learned predictions are applied to program executions in order to save resources. The three books will have the following subtopics: Volume 1: Machine Learning under Resource Constraints - Fundamentals Volume 2: Machine Learning and Physics under Resource Constraints - Discovery Volume 3: Machine Learning under Resource Constraints - Applications Volume 1 establishes the foundations of this new field (Machine Learning under Resource Constraints). It goes through all the steps from data collection, their summary and clustering, to the different aspects of resource-aware learning, i.e., hardware, memory, energy, and communication awareness. Several machine learning methods are inspected with respect to their resource requirements and how to enhance their scalability on diverse computing architectures ranging from embedded systems to large computing clusters.




Computational Intelligence for Green Cloud Computing and Digital Waste Management


Book Description

In the digital age, the relentless growth of data centers and cloud computing has given rise to a pressing dilemma. The power consumption of these facilities is spiraling out of control, emitting massive amounts of carbon dioxide, and contributing to the ever-increasing threat of global warming. Studies show that data centers alone are responsible for nearly eighty million metric tons of CO2 emissions worldwide, and this figure is poised to skyrocket to a staggering 8000 TWh by 2030 unless we revolutionize our approach to computing resource management. The root of this problem lies in inefficient resource allocation within cloud environments, as service providers often over-provision computing resources to avoid Service Level Agreement (SLA) violations, leading to both underutilization of resources and a significant increase in energy consumption. Computational Intelligence for Green Cloud Computing and Digital Waste Management stands as a beacon of hope in the face of the environmental and technological challenges we face. It introduces the concept of green computing, dedicated to creating an eco-friendly computing environment. The book explores innovative, intelligent resource management methods that can significantly reduce the power consumption of data centers. From machine learning and deep learning solutions to green virtualization technologies, this comprehensive guide explores innovative approaches to address the pressing challenges of green computing. Whether you are an educator teaching about green computing, an environmentalist seeking sustainability solutions, an industry professional navigating the digital landscape, a resolute researcher, or simply someone intrigued by the intersection of technology and sustainability, this book offers an indispensable resource.




Embedded Computer Systems: Architectures, Modeling, and Simulation


Book Description

This book constitutes the proceedings of the 21st International Conference on Embedded Computer Systems: Architectures, Modeling, and Simulation, SAMOS 2021, which took place in July 2021. Due to COVID-19 pandemic the conference was held virtually. The 17 full papers presented in this volume were carefully reviewed and selected from 45 submissions. The papers are organized in topics as follows: simulation and design space exploration; the 3Cs - Cache, Cluster and Cloud; heterogeneous SoC; novel CPU architectures and applications; dataflow; innovative architectures and tools for security; next generation computing; insights from negative results.




Unlocking Artificial Intelligence


Book Description

This open access book provides a state-of-the-art overview of current machine learning research and its exploitation in various application areas. It has become apparent that the deep integration of artificial intelligence (AI) methods in products and services is essential for companies to stay competitive. The use of AI allows large volumes of data to be analyzed, patterns and trends to be identified, and well-founded decisions to be made on an informative basis. It also enables the optimization of workflows, the automation of processes and the development of new services, thus creating potential for new business models and significant competitive advantages. The book is divided in two main parts: First, in a theoretically oriented part, various AI/ML-related approaches like automated machine learning, sequence-based learning, deep learning, learning from experience and data, and process-aware learning are explained. In a second part, various applications are presented that benefit from the exploitation of recent research results. These include autonomous systems, indoor localization, medical applications, energy supply and networks, logistics networks, traffic control, image processing, and IoT applications. Overall, the book offers professionals and applied researchers an excellent overview of current exploitations, approaches, and challenges of AI/ML-related research.




Large-Scale Annotation of Biomedical Data and Expert Label Synthesis and Hardware Aware Learning for Medical Imaging and Computer Assisted Intervention


Book Description

This book constitutes the refereed joint proceedings of the 4th International Workshop on Large-Scale Annotation of Biomedical Data and Expert Label Synthesis, LABELS 2019, the First International Workshop on Hardware Aware Learning for Medical Imaging and Computer Assisted Intervention, HAL-MICCAI 2019, and the Second International Workshop on Correction of Brainshift with Intra-Operative Ultrasound, CuRIOUS 2019, held in conjunction with the 22nd International Conference on Medical Imaging and Computer-Assisted Intervention, MICCAI 2019, in Shenzhen, China, in October 2019. The 8 papers presented at LABELS 2019, the 5 papers presented at HAL-MICCAI 2019, and the 3 papers presented at CuRIOUS 2019 were carefully reviewed and selected from numerous submissions. The LABELS papers present a variety of approaches for dealing with a limited number of labels, from semi-supervised learning to crowdsourcing. The HAL-MICCAI papers cover a wide set of hardware applications in medical problems, including medical image segmentation, electron tomography, pneumonia detection, etc. The CuRIOUS papers provide a snapshot of the current progress in the field through extended discussions and provide researchers an opportunity to characterize their image registration methods on newly released standardized datasets of iUS-guided brain tumor resection.




Deep Learning Systems


Book Description

This book describes deep learning systems: the algorithms, compilers, and processor components to efficiently train and deploy deep learning models for commercial applications. The exponential growth in computational power is slowing at a time when the amount of compute consumed by state-of-the-art deep learning (DL) workloads is rapidly growing. Model size, serving latency, and power constraints are a significant challenge in the deployment of DL models for many applications. Therefore, it is imperative to codesign algorithms, compilers, and hardware to accelerate advances in this field with holistic system-level and algorithm solutions that improve performance, power, and efficiency. Advancing DL systems generally involves three types of engineers: (1) data scientists that utilize and develop DL algorithms in partnership with domain experts, such as medical, economic, or climate scientists; (2) hardware designers that develop specialized hardware to accelerate the components in the DL models; and (3) performance and compiler engineers that optimize software to run more efficiently on a given hardware. Hardware engineers should be aware of the characteristics and components of production and academic models likely to be adopted by industry to guide design decisions impacting future hardware. Data scientists should be aware of deployment platform constraints when designing models. Performance engineers should support optimizations across diverse models, libraries, and hardware targets. The purpose of this book is to provide a solid understanding of (1) the design, training, and applications of DL algorithms in industry; (2) the compiler techniques to map deep learning code to hardware targets; and (3) the critical hardware features that accelerate DL systems. This book aims to facilitate co-innovation for the advancement of DL systems. It is written for engineers working in one or more of these areas who seek to understand the entire system stack in order to better collaborate with engineers working in other parts of the system stack. The book details advancements and adoption of DL models in industry, explains the training and deployment process, describes the essential hardware architectural features needed for today's and future models, and details advances in DL compilers to efficiently execute algorithms across various hardware targets. Unique in this book is the holistic exposition of the entire DL system stack, the emphasis on commercial applications, and the practical techniques to design models and accelerate their performance. The author is fortunate to work with hardware, software, data scientist, and research teams across many high-technology companies with hyperscale data centers. These companies employ many of the examples and methods provided throughout the book.