Interpretable Machine Learning


Book Description

This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.




Explainable and Interpretable Models in Computer Vision and Machine Learning


Book Description

This book compiles leading research on the development of explainable and interpretable machine learning methods in the context of computer vision and machine learning. Research progress in computer vision and pattern recognition has led to a variety of modeling techniques with almost human-like performance. Although these models have obtained astounding results, they are limited in their explainability and interpretability: what is the rationale behind the decision made? what in the model structure explains its functioning? Hence, while good performance is a critical required characteristic for learning machines, explainability and interpretability capabilities are needed to take learning machines to the next step to include them in decision support systems involving human supervision. This book, written by leading international researchers, addresses key topics of explainability and interpretability, including the following: · Evaluation and Generalization in Interpretable Machine Learning · Explanation Methods in Deep Learning · Learning Functional Causal Models with Generative Neural Networks · Learning Interpreatable Rules for Multi-Label Classification · Structuring Neural Networks for More Explainable Predictions · Generating Post Hoc Rationales of Deep Visual Classification Decisions · Ensembling Visual Explanations · Explainable Deep Driving by Visualizing Causal Attention · Interdisciplinary Perspective on Algorithmic Job Candidate Search · Multimodal Personality Trait Analysis for Explainable Modeling of Job Interview Decisions · Inherent Explainability Pattern Theory-based Video Event Interpretations




Industry 4.0 for the Built Environment


Book Description

This book discusses how the role of traditional construction professional is changing, providing a useful guide for practitioners who would like to upskill themselves. Lately, core concepts and methodologies for the Built Environment are presented providing definitions and applications on Building Information Modelling, Computational Design, Artificial Intelligence, Big Data, Cloud Computing, Data Analytics and Visualization, Lean Construction, Advanced Project Management, Sustainability, Geographical Information Systems, Advanced Business Models, Disaster Management, Quality Management, Health and Safety and Legal prospective. The book also shows the latest technologies for the Built Environment including Digital Twins, Reality Capture, Extended Reality, Gamification, Computational Construction and Manufacturing, Structural Health Monitoring, Smart Transaction and Cybersecurity. Trends in soft skills for the Built Environment are presented covering Digital Working, Communication, Self and Relationship Management skills and Critical thinking. The book is dedicated to professionals who would like to enhance their understanding and capabilities to operate in the Industry 4.0 for the Built Environment having a holistic and comprehensive overview.




Explainable AI: Interpreting, Explaining and Visualizing Deep Learning


Book Description

The development of “intelligent” systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to “intelligent” machines. For sensitive tasks involving critical infrastructures and affecting human well-being or health, it is crucial to limit the possibility of improper, non-robust and unsafe decisions and actions. Before deploying an AI system, we see a strong need to validate its behavior, and thus establish guarantees that it will continue to perform as expected when deployed in a real-world environment. In pursuit of that objective, ways for humans to verify the agreement between the AI decision structure and their own ground-truth knowledge have been explored. Explainable AI (XAI) has developed as a subfield of AI, focused on exposing complex AI models to humans in a systematic and interpretable manner. The 22 chapters included in this book provide a timely snapshot of algorithms, theory, and applications of interpretable and explainable AI and AI techniques that have been proposed recently reflecting the current discourse in this field and providing directions of future development. The book is organized in six parts: towards AI transparency; methods for interpreting AI systems; explaining the decisions of AI systems; evaluating interpretability and explanations; applications of explainable AI; and software for explainable AI.




Handbook On Big Data And Machine Learning In The Physical Sciences (In 2 Volumes)


Book Description

This compendium provides a comprehensive collection of the emergent applications of big data, machine learning, and artificial intelligence technologies to present day physical sciences ranging from materials theory and imaging to predictive synthesis and automated research. This area of research is among the most rapidly developing in the last several years in areas spanning materials science, chemistry, and condensed matter physics.Written by world renowned researchers, the compilation of two authoritative volumes provides a distinct summary of the modern advances in instrument — driven data generation and analytics, establishing the links between the big data and predictive theories, and outlining the emerging field of data and physics-driven predictive and autonomous systems.




Interpretable Machine Learning with Python


Book Description

A deep and detailed dive into the key aspects and challenges of machine learning interpretability, complete with the know-how on how to overcome and leverage them to build fairer, safer, and more reliable models Key Features Learn how to extract easy-to-understand insights from any machine learning model Become well-versed with interpretability techniques to build fairer, safer, and more reliable models Mitigate risks in AI systems before they have broader implications by learning how to debug black-box models Book DescriptionDo you want to gain a deeper understanding of your models and better mitigate poor prediction risks associated with machine learning interpretation? If so, then Interpretable Machine Learning with Python deserves a place on your bookshelf. We’ll be starting off with the fundamentals of interpretability, its relevance in business, and exploring its key aspects and challenges. As you progress through the chapters, you'll then focus on how white-box models work, compare them to black-box and glass-box models, and examine their trade-off. You’ll also get you up to speed with a vast array of interpretation methods, also known as Explainable AI (XAI) methods, and how to apply them to different use cases, be it for classification or regression, for tabular, time-series, image or text. In addition to the step-by-step code, this book will also help you interpret model outcomes using examples. You’ll get hands-on with tuning models and training data for interpretability by reducing complexity, mitigating bias, placing guardrails, and enhancing reliability. The methods you’ll explore here range from state-of-the-art feature selection and dataset debiasing methods to monotonic constraints and adversarial retraining. By the end of this book, you'll be able to understand ML models better and enhance them through interpretability tuning. What you will learn Recognize the importance of interpretability in business Study models that are intrinsically interpretable such as linear models, decision trees, and Naïve Bayes Become well-versed in interpreting models with model-agnostic methods Visualize how an image classifier works and what it learns Understand how to mitigate the influence of bias in datasets Discover how to make models more reliable with adversarial robustness Use monotonic constraints to make fairer and safer models Who this book is for This book is primarily written for data scientists, machine learning developers, and data stewards who find themselves under increasing pressures to explain the workings of AI systems, their impacts on decision making, and how they identify and manage bias. It’s also a useful resource for self-taught ML enthusiasts and beginners who want to go deeper into the subject matter, though a solid grasp on the Python programming language and ML fundamentals is needed to follow along.




TinyML


Book Description

Deep learning networks are getting smaller. Much smaller. The Google Assistant team can detect words with a model just 14 kilobytes in size—small enough to run on a microcontroller. With this practical book you’ll enter the field of TinyML, where deep learning and embedded systems combine to make astounding things possible with tiny devices. Pete Warden and Daniel Situnayake explain how you can train models small enough to fit into any environment. Ideal for software and hardware developers who want to build embedded systems using machine learning, this guide walks you through creating a series of TinyML projects, step-by-step. No machine learning or microcontroller experience is necessary. Build a speech recognizer, a camera that detects people, and a magic wand that responds to gestures Work with Arduino and ultra-low-power microcontrollers Learn the essentials of ML and how to train your own models Train models to understand audio, image, and accelerometer data Explore TensorFlow Lite for Microcontrollers, Google’s toolkit for TinyML Debug applications and provide safeguards for privacy and security Optimize latency, energy usage, and model and binary size




Applications of Big Data in Large- and Small-Scale Systems


Book Description

With new technologies, such as computer vision, internet of things, mobile computing, e-governance and e-commerce, and wide applications of social media, organizations generate a huge volume of data and at a much faster rate than several years ago. Big data in large-/small-scale systems, characterized by high volume, diversity, and velocity, increasingly drives decision making and is changing the landscape of business intelligence. From governments to private organizations, from communities to individuals, all areas are being affected by this shift. There is a high demand for big data analytics that offer insights for computing efficiency, knowledge discovery, problem solving, and event prediction. To handle this demand and this increase in big data, there needs to be research on innovative and optimized machine learning algorithms in both large- and small-scale systems. Applications of Big Data in Large- and Small-Scale Systems includes state-of-the-art research findings on the latest development, up-to-date issues, and challenges in the field of big data and presents the latest innovative and intelligent applications related to big data. This book encompasses big data in various multidisciplinary fields from the medical field to agriculture, business research, and smart cities. While highlighting topics including machine learning, cloud computing, data visualization, and more, this book is a valuable reference tool for computer scientists, data scientists and analysts, engineers, practitioners, stakeholders, researchers, academicians, and students interested in the versatile and innovative use of big data in both large-scale and small-scale systems.




Explainable, Interpretable, and Transparent AI Systems


Book Description

Transparent Artificial Intelligence (AI) systems facilitate understanding of the decision-making process and provide opportunities in various aspects of explaining AI models. This book provides up-to-date information on the latest advancements in the field of explainable AI, which is a critical requirement of AI, Machine Learning (ML), and Deep Learning (DL) models. It provides examples, case studies, latest techniques, and applications from domains such as healthcare, finance, and network security. It also covers open-source interpretable tool kits so that practitioners can use them in their domains. Features: Presents a clear focus on the application of explainable AI systems while tackling important issues of “interpretability” and “transparency”. Reviews adept handling with respect to existing software and evaluation issues of interpretability. Provides insights into simple interpretable models such as decision trees, decision rules, and linear regression. Focuses on interpreting black box models like feature importance and accumulated local effects. Discusses capabilities of explainability and interpretability. This book is aimed at graduate students and professionals in computer engineering and networking communications.




Knowledge Graphs for eXplainable Artificial Intelligence: Foundations, Applications and Challenges


Book Description

The latest advances in Artificial Intelligence and (deep) Machine Learning in particular revealed a major drawback of modern intelligent systems, namely the inability to explain their decisions in a way that humans can easily understand. While eXplainable AI rapidly became an active area of research in response to this need for improved understandability and trustworthiness, the field of Knowledge Representation and Reasoning (KRR) has on the other hand a long-standing tradition in managing information in a symbolic, human-understandable form. This book provides the first comprehensive collection of research contributions on the role of knowledge graphs for eXplainable AI (KG4XAI), and the papers included here present academic and industrial research focused on the theory, methods and implementations of AI systems that use structured knowledge to generate reliable explanations. Introductory material on knowledge graphs is included for those readers with only a minimal background in the field, as well as specific chapters devoted to advanced methods, applications and case-studies that use knowledge graphs as a part of knowledge-based, explainable systems (KBX-systems). The final chapters explore current challenges and future research directions in the area of knowledge graphs for eXplainable AI. The book not only provides a scholarly, state-of-the-art overview of research in this subject area, but also fosters the hybrid combination of symbolic and subsymbolic AI methods, and will be of interest to all those working in the field.