Interpretability for Industry 4.0 : Statistical and Machine Learning Approaches


Book Description

This volume provides readers with a compact, stimulating and multifaceted introduction to interpretability, a key issue for developing insightful statistical and machine learning approaches as well as for communicating modelling results in business and industry. Different views in the context of Industry 4.0 are offered in connection with the concepts of explainability of machine learning tools, generalizability of model outputs and sensitivity analysis. Moreover, the book explores the integration of Artificial Intelligence and robust analysis of variance for big data mining and monitoring in Additive Manufacturing, and sheds new light on interpretability via random forests and flexible generalized additive models together with related software resources and real-world examples.




Interpretable Machine Learning


Book Description

This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.




Interpretable Machine Learning and Generative Modeling with Mixed Tabular Data


Book Description

Explainable artificial intelligence or interpretable machine learning techniques aim to shed light on the behavior of opaque machine learning algorithms, yet often fail to acknowledge the challenges real-world data imposes on the task. Specifically, the fact that empirical tabular datasets may consist of both continuous and categorical features (mixed data) and typically exhibit dependency structures is frequently overlooked. This work uses a statistical perspective to illuminate the far-reaching implications of mixed data and dependency structures for interpretability in machine learning. Several interpretability methods are advanced with a particular focus on this kind of data, evaluating their performance on simulated and real data sets. Further, this cumulative thesis emphasizes that generating synthetic data is a crucial subroutine for many interpretability methods. Therefore, this thesis also advances methodology in generative modeling concerning mixed tabular data, presenting a tree-based approach for density estimation and data generation, accompanied by a user-friendly software implementation in the Python programming language.







Interpretable Statistical Learning


Book Description

Interpretability of machine learning models is important in critical applications to attain trust of users. Despite their strong performance, black-box machine learning models often meet resistance in usage, especially in areas such as economics, social science, healthcare industry, and administrative decision making. This dissertation explores methods to improve 'human interpretability' for both supervised and unsupervised machine learning. I approach this topic by building statistical models with relatively low complexity and developing post-hoc model-agnostic tools. This dissertation consists of three projects. In the first project, we propose a new method to estimate a mixture of linear models (MLM) for regression or classification that is relatively easy to interpret. We use DNN as a proxy of the optimal prediction function so that MLM can be effectively estimated. We propose visualization methods and quantitative approaches to interpret the predictor by MLM. Experiments show that the new method allows us to trade-off interpretability and accuracy. MLM estimated under the guidance of a trained DNN fills the gap between a highly explainable linear statistical model and a highly accurate but difficult to interpret predictor. In the second project, we develop a new block-wise variable selection method for clustering by exploiting the latent states of the hidden Markov model on variable blocks or the Gaussian mixture model. Specifically, the variable blocks are formed by depth-first-search on a dendrogram created based on the mutual information between any pair of variables. It is demonstrated that the latent states of the variable blocks together with the mixture model parameters can represent the original data effectively and much more compactly. We thus cluster the data using the latent states and select variables according to the relationship between the states and the clusters. As true class labels are unknown in the unsupervised setting, we first generate more refined clusters, namely, semi-clusters, for variable selection and then determine the final clusters based on the dimension reduced data. The new method increases the interpretability of high-dimensional clustering by effectively reducing the model complexity and selecting variables while retains the comparable clustering accuracy to other widely used methods. In the third project, we propose a new framework to interpret and validate clustering results for any baseline methods. We exploit the optimal transport alignment and the bootstrapping method to quantify the variation of clustering results at the levels of both overall partitions and individual clusters. Set relationships between clusters such as one-to-one match, split, and merge can be revealed. A covering point set for each cluster, a concept kin to the confidence interval, is proposed. The tools we have developed here will help understand the model behavior of the baseline clustering method. Experimental results on both simulated and real datasets are provided. The corresponding R package OTclust is available on CRAN.







Explainable and Interpretable Models in Computer Vision and Machine Learning


Book Description

This book compiles leading research on the development of explainable and interpretable machine learning methods in the context of computer vision and machine learning. Research progress in computer vision and pattern recognition has led to a variety of modeling techniques with almost human-like performance. Although these models have obtained astounding results, they are limited in their explainability and interpretability: what is the rationale behind the decision made? what in the model structure explains its functioning? Hence, while good performance is a critical required characteristic for learning machines, explainability and interpretability capabilities are needed to take learning machines to the next step to include them in decision support systems involving human supervision. This book, written by leading international researchers, addresses key topics of explainability and interpretability, including the following: · Evaluation and Generalization in Interpretable Machine Learning · Explanation Methods in Deep Learning · Learning Functional Causal Models with Generative Neural Networks · Learning Interpreatable Rules for Multi-Label Classification · Structuring Neural Networks for More Explainable Predictions · Generating Post Hoc Rationales of Deep Visual Classification Decisions · Ensembling Visual Explanations · Explainable Deep Driving by Visualizing Causal Attention · Interdisciplinary Perspective on Algorithmic Job Candidate Search · Multimodal Personality Trait Analysis for Explainable Modeling of Job Interview Decisions · Inherent Explainability Pattern Theory-based Video Event Interpretations




Automated Machine Learning


Book Description

This open access book presents the first comprehensive overview of general methods in Automated Machine Learning (AutoML), collects descriptions of existing systems based on these methods, and discusses the first series of international challenges of AutoML systems. The recent success of commercial ML applications and the rapid growth of the field has created a high demand for off-the-shelf ML methods that can be used easily and without expert knowledge. However, many of the recent machine learning successes crucially rely on human experts, who manually select appropriate ML architectures (deep learning architectures or more traditional ML workflows) and their hyperparameters. To overcome this problem, the field of AutoML targets a progressive automation of machine learning, based on principles from optimization and machine learning itself. This book serves as a point of entry into this quickly-developing field for researchers and advanced students alike, as well as providing a reference for practitioners aiming to use AutoML in their work.




An Introduction to Statistical Learning


Book Description

An Introduction to Statistical Learning provides an accessible overview of the field of statistical learning, an essential toolset for making sense of the vast and complex data sets that have emerged in fields ranging from biology to finance, marketing, and astrophysics in the past twenty years. This book presents some of the most important modeling and prediction techniques, along with relevant applications. Topics include linear regression, classification, resampling methods, shrinkage approaches, tree-based methods, support vector machines, clustering, deep learning, survival analysis, multiple testing, and more. Color graphics and real-world examples are used to illustrate the methods presented. This book is targeted at statisticians and non-statisticians alike, who wish to use cutting-edge statistical learning techniques to analyze their data. Four of the authors co-wrote An Introduction to Statistical Learning, With Applications in R (ISLR), which has become a mainstay of undergraduate and graduate classrooms worldwide, as well as an important reference book for data scientists. One of the keys to its success was that each chapter contains a tutorial on implementing the analyses and methods presented in the R scientific computing environment. However, in recent years Python has become a popular language for data science, and there has been increasing demand for a Python-based alternative to ISLR. Hence, this book (ISLP) covers the same materials as ISLR but with labs implemented in Python. These labs will be useful both for Python novices, as well as experienced users.




Interpretable Machine Learning with Python


Book Description

A deep and detailed dive into the key aspects and challenges of machine learning interpretability, complete with the know-how on how to overcome and leverage them to build fairer, safer, and more reliable models Key Features Learn how to extract easy-to-understand insights from any machine learning model Become well-versed with interpretability techniques to build fairer, safer, and more reliable models Mitigate risks in AI systems before they have broader implications by learning how to debug black-box models Book DescriptionDo you want to gain a deeper understanding of your models and better mitigate poor prediction risks associated with machine learning interpretation? If so, then Interpretable Machine Learning with Python deserves a place on your bookshelf. We’ll be starting off with the fundamentals of interpretability, its relevance in business, and exploring its key aspects and challenges. As you progress through the chapters, you'll then focus on how white-box models work, compare them to black-box and glass-box models, and examine their trade-off. You’ll also get you up to speed with a vast array of interpretation methods, also known as Explainable AI (XAI) methods, and how to apply them to different use cases, be it for classification or regression, for tabular, time-series, image or text. In addition to the step-by-step code, this book will also help you interpret model outcomes using examples. You’ll get hands-on with tuning models and training data for interpretability by reducing complexity, mitigating bias, placing guardrails, and enhancing reliability. The methods you’ll explore here range from state-of-the-art feature selection and dataset debiasing methods to monotonic constraints and adversarial retraining. By the end of this book, you'll be able to understand ML models better and enhance them through interpretability tuning. What you will learn Recognize the importance of interpretability in business Study models that are intrinsically interpretable such as linear models, decision trees, and Naïve Bayes Become well-versed in interpreting models with model-agnostic methods Visualize how an image classifier works and what it learns Understand how to mitigate the influence of bias in datasets Discover how to make models more reliable with adversarial robustness Use monotonic constraints to make fairer and safer models Who this book is for This book is primarily written for data scientists, machine learning developers, and data stewards who find themselves under increasing pressures to explain the workings of AI systems, their impacts on decision making, and how they identify and manage bias. It’s also a useful resource for self-taught ML enthusiasts and beginners who want to go deeper into the subject matter, though a solid grasp on the Python programming language and ML fundamentals is needed to follow along.