Explainable Machine Learning Models and Architectures


Book Description

EXPLAINABLE MACHINE LEARNING MODELS AND ARCHITECTURES This cutting-edge new volume covers the hardware architecture implementation, the software implementation approach, and the efficient hardware of machine learning applications. Machine learning and deep learning modules are now an integral part of many smart and automated systems where signal processing is performed at different levels. Signal processing in the form of text, images, or video needs large data computational operations at the desired data rate and accuracy. Large data requires more use of integrated circuit (IC) area with embedded bulk memories that further lead to more IC area. Trade-offs between power consumption, delay and IC area are always a concern of designers and researchers. New hardware architectures and accelerators are needed to explore and experiment with efficient machine-learning models. Many real-time applications like the processing of biomedical data in healthcare, smart transportation, satellite image analysis, and IoT-enabled systems have a lot of scope for improvements in terms of accuracy, speed, computational powers, and overall power consumption. This book deals with the efficient machine and deep learning models that support high-speed processors with reconfigurable architectures like graphic processing units (GPUs) and field programmable gate arrays (FPGAs), or any hybrid system. Whether for the veteran engineer or scientist working in the field or laboratory, or the student or academic, this is a must-have for any library.




Interpretable Machine Learning


Book Description

This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.




Explainable Machine Learning Models and Architectures


Book Description

EXPLAINABLE MACHINE LEARNING MODELS AND ARCHITECTURES This cutting-edge new volume covers the hardware architecture implementation, the software implementation approach, and the efficient hardware of machine learning applications. Machine learning and deep learning modules are now an integral part of many smart and automated systems where signal processing is performed at different levels. Signal processing in the form of text, images, or video needs large data computational operations at the desired data rate and accuracy. Large data requires more use of integrated circuit (IC) area with embedded bulk memories that further lead to more IC area. Trade-offs between power consumption, delay and IC area are always a concern of designers and researchers. New hardware architectures and accelerators are needed to explore and experiment with efficient machine-learning models. Many real-time applications like the processing of biomedical data in healthcare, smart transportation, satellite image analysis, and IoT-enabled systems have a lot of scope for improvements in terms of accuracy, speed, computational powers, and overall power consumption. This book deals with the efficient machine and deep learning models that support high-speed processors with reconfigurable architectures like graphic processing units (GPUs) and field programmable gate arrays (FPGAs), or any hybrid system. Whether for the veteran engineer or scientist working in the field or laboratory, or the student or academic, this is a must-have for any library.




Explainable AI: Interpreting, Explaining and Visualizing Deep Learning


Book Description

The development of “intelligent” systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to “intelligent” machines. For sensitive tasks involving critical infrastructures and affecting human well-being or health, it is crucial to limit the possibility of improper, non-robust and unsafe decisions and actions. Before deploying an AI system, we see a strong need to validate its behavior, and thus establish guarantees that it will continue to perform as expected when deployed in a real-world environment. In pursuit of that objective, ways for humans to verify the agreement between the AI decision structure and their own ground-truth knowledge have been explored. Explainable AI (XAI) has developed as a subfield of AI, focused on exposing complex AI models to humans in a systematic and interpretable manner. The 22 chapters included in this book provide a timely snapshot of algorithms, theory, and applications of interpretable and explainable AI and AI techniques that have been proposed recently reflecting the current discourse in this field and providing directions of future development. The book is organized in six parts: towards AI transparency; methods for interpreting AI systems; explaining the decisions of AI systems; evaluating interpretability and explanations; applications of explainable AI; and software for explainable AI.




Explainable and Interpretable Models in Computer Vision and Machine Learning


Book Description

This book compiles leading research on the development of explainable and interpretable machine learning methods in the context of computer vision and machine learning. Research progress in computer vision and pattern recognition has led to a variety of modeling techniques with almost human-like performance. Although these models have obtained astounding results, they are limited in their explainability and interpretability: what is the rationale behind the decision made? what in the model structure explains its functioning? Hence, while good performance is a critical required characteristic for learning machines, explainability and interpretability capabilities are needed to take learning machines to the next step to include them in decision support systems involving human supervision. This book, written by leading international researchers, addresses key topics of explainability and interpretability, including the following: · Evaluation and Generalization in Interpretable Machine Learning · Explanation Methods in Deep Learning · Learning Functional Causal Models with Generative Neural Networks · Learning Interpreatable Rules for Multi-Label Classification · Structuring Neural Networks for More Explainable Predictions · Generating Post Hoc Rationales of Deep Visual Classification Decisions · Ensembling Visual Explanations · Explainable Deep Driving by Visualizing Causal Attention · Interdisciplinary Perspective on Algorithmic Job Candidate Search · Multimodal Personality Trait Analysis for Explainable Modeling of Job Interview Decisions · Inherent Explainability Pattern Theory-based Video Event Interpretations




Knowledge Graphs for eXplainable Artificial Intelligence: Foundations, Applications and Challenges


Book Description

The latest advances in Artificial Intelligence and (deep) Machine Learning in particular revealed a major drawback of modern intelligent systems, namely the inability to explain their decisions in a way that humans can easily understand. While eXplainable AI rapidly became an active area of research in response to this need for improved understandability and trustworthiness, the field of Knowledge Representation and Reasoning (KRR) has on the other hand a long-standing tradition in managing information in a symbolic, human-understandable form. This book provides the first comprehensive collection of research contributions on the role of knowledge graphs for eXplainable AI (KG4XAI), and the papers included here present academic and industrial research focused on the theory, methods and implementations of AI systems that use structured knowledge to generate reliable explanations. Introductory material on knowledge graphs is included for those readers with only a minimal background in the field, as well as specific chapters devoted to advanced methods, applications and case-studies that use knowledge graphs as a part of knowledge-based, explainable systems (KBX-systems). The final chapters explore current challenges and future research directions in the area of knowledge graphs for eXplainable AI. The book not only provides a scholarly, state-of-the-art overview of research in this subject area, but also fosters the hybrid combination of symbolic and subsymbolic AI methods, and will be of interest to all those working in the field.




Deep Learning in Medical Image Analysis


Book Description

This book presents cutting-edge research and applications of deep learning in a broad range of medical imaging scenarios, such as computer-aided diagnosis, image segmentation, tissue recognition and classification, and other areas of medical and healthcare problems. Each of its chapters covers a topic in depth, ranging from medical image synthesis and techniques for muskuloskeletal analysis to diagnostic tools for breast lesions on digital mammograms and glaucoma on retinal fundus images. It also provides an overview of deep learning in medical image analysis and highlights issues and challenges encountered by researchers and clinicians, surveying and discussing practical approaches in general and in the context of specific problems. Academics, clinical and industry researchers, as well as young researchers and graduate students in medical imaging, computer-aided-diagnosis, biomedical engineering and computer vision will find this book a great reference and very useful learning resource.







Machine Learning Design Patterns


Book Description

The design patterns in this book capture best practices and solutions to recurring problems in machine learning. The authors, three Google engineers, catalog proven methods to help data scientists tackle common problems throughout the ML process. These design patterns codify the experience of hundreds of experts into straightforward, approachable advice. In this book, you will find detailed explanations of 30 patterns for data and problem representation, operationalization, repeatability, reproducibility, flexibility, explainability, and fairness. Each pattern includes a description of the problem, a variety of potential solutions, and recommendations for choosing the best technique for your situation. You'll learn how to: Identify and mitigate common challenges when training, evaluating, and deploying ML models Represent data for different ML model types, including embeddings, feature crosses, and more Choose the right model type for specific problems Build a robust training loop that uses checkpoints, distribution strategy, and hyperparameter tuning Deploy scalable ML systems that you can retrain and update to reflect new data Interpret model predictions for stakeholders and ensure models are treating users fairly




Explainable Artificial Intelligence: An Introduction to Interpretable Machine Learning


Book Description

This book is written both for readers entering the field, and for practitioners with a background in AI and an interest in developing real-world applications. The book is a great resource for practitioners and researchers in both industry and academia, and the discussed case studies and associated material can serve as inspiration for a variety of projects and hands-on assignments in a classroom setting. I will certainly keep this book as a personal resource for the courses I teach, and strongly recommend it to my students. --Dr. Carlotta Domeniconi, Associate Professor, Computer Science Department, GMU This book offers a curriculum for introducing interpretability to machine learning at every stage. The authors provide compelling examples that a core teaching practice like leading interpretive discussions can be taught and learned by teachers and sustained effort. And what better way to strengthen the quality of AI and Machine learning outcomes. I hope that this book will become a primer for teachers, data Science educators, and ML developers, and together we practice the art of interpretive machine learning. --Anusha Dandapani, Chief Data and Analytics Officer, UNICC and Adjunct Faculty, NYU This is a wonderful book! I’m pleased that the next generation of scientists will finally be able to learn this important topic. This is the first book I’ve seen that has up-to-date and well-rounded coverage. Thank you to the authors! --Dr. Cynthia Rudin, Professor of Computer Science, Electrical and Computer Engineering, Statistical Science, and Biostatistics & Bioinformatics Literature on Explainable AI has up until now been relatively scarce and featured mainly mainstream algorithms like SHAP and LIME. This book has closed this gap by providing an extremely broad review of various algorithms proposed in the scientific circles over the previous 5-10 years. This book is a great guide to anyone who is new to the field of XAI or is already familiar with the field and is willing to expand their knowledge. A comprehensive review of the state-of-the-art Explainable AI methods starting from visualization, interpretable methods, local and global explanations, time series methods, and finishing with deep learning provides an unparalleled source of information currently unavailable anywhere else. Additionally, notebooks with vivid examples are a great supplement that makes the book even more attractive for practitioners of any level. Overall, the authors provide readers with an enormous breadth of coverage without losing sight of practical aspects, which makes this book truly unique and a great addition to the library of any data scientist. Dr. Andrey Sharapov, Product Data Scientist, Explainable AI Expert and Speaker, Founder of Explainable AI-XAI Group