Towards Distribution-free Interpretation, Inference and Network Estimation


Book Description

In the era of AI, statistical or machine learning methods towards distribution-free assumptions are becoming increasingly important due to the growing amount of data that is being collected and analyzed. Traditional parametric methods may not always be appropriate or may lead to model mis-specification and inaccurate results when dealing with large or complex data sets. Besides, as specific distributional assumptions or parametric modeling are removed, the challenge of model interpretation and prediction inference arises and has been currently at the forefront of research efforts. One problem of our interests in this regard is non-parametric or semi-parametric network estimation for data that are not independent. Specifically, influence network estimation from a multi-variate point process or time series data is a problem of fundamental importance. Prior work has focused on parametric approaches that require a known parametric model, which makes estimation procedures less robust to model mis-specification, non-linearities and heterogeneities. In Chapter 2, we develop a semi-parametric approach based on the monotone single-index multi-variate autoregressive model (SIMAM) which addresses these challenges. In particular, rather than using standard parametric approaches, we use the monotone single index model (SIM) for network estimation. We provide theoretical guarantees for dependent data, and an alternating projected gradient descent algorithm. Significantly we achieve rates of the form O(T^{-1/3} \sqrt{s\log(TM)}) (optimal in the independent design case) where s is {he number of edges in the influence network that indicates the sparsity level, M is the number of actors and T is the number of time points. In addition, we demonstrate the performance of SIMAM both on simulated data and two real data examples, and show it outperforms state-of-the-art parametric methods both in terms of prediction and network estimation. Another aspect important for distribution-free or model-free learning is the interpretation, i.e. to make the complicated non-parametric predictive models explainable. A number of model-agnostic methods for measuring variable importance (VI) have emerged in recent times, which assess the difference in predictive power between a full model trained on all variables and a reduced model that omits the variable(s) of interest. However, these methods typically encounter a bottleneck when estimating the reduced model for each variable or subset of variables, which is both costly and lacks theoretical guarantees. To address this problem, Chapter 3 proposes an efficient and adaptable approach for approximating the reduced model while ensuring important inferential guarantees. Specifically, we replace the need for fully retraining a wide neural network with a linearization that is initiated using the full model parameters. By including a ridge-like penalty to make the problem convex, we establish that our method can estimate the variable importance measure with an error rate of O({1}/{\sqrt{n}), where n represents the number of training samples, provided that the ridge penalty parameter is adequately large. Furthermore, we demonstrate that our estimator is asymptotically normal, enabling us to provide confidence bounds for the VI estimates. Finally, we demonstrate the method's speed and accuracy under different data-generating regimes and showcase its applicability in a real-world seasonal climate forecasting example. In addition to semi-parametric network estimation and fast estimation of variable importance for interpretation, an efficient method for prediction inference without specific distributional assumptions on the data is of our interest as well. In Chapter 4, we present a novel, computationally-efficient algorithm for predictive inference (PI) that requires no distributional assumptions in the data and can be computed faster than existing bootstrap-type methods for neural networks. Specifically, if there are $n$ training samples, bootstrap methods require training a model on each of the n subsamples of size n-1; for large models like neural networks, this process can be computationally prohibitive. In contrast, the proposed method trains one neural network on the full dataset with ([epsilon], [delta]) -differential privacy (DP) and then approximates each leave-one-out model efficiently using a linear approximation around the neural network estimate. With exchangeable data, we prove that our approach has a rigorous coverage guarantee that depends on the preset privacy parameters and the stability of the neural network, regardless of the data distribution. Simulations and experiments on real data demonstrate that our method satisfies the coverage guarantees with substantially reduced computation compared to bootstrap methods.




Towards Distribution-free Interpretation, Inference and Network Estimation


Book Description

In the era of AI, statistical or machine learning methods towards distribution-free assumptions are becoming increasingly important due to the growing amount of data that is being collected and analyzed. Traditional parametric methods may not always be appropriate or may lead to model mis-specification and inaccurate results when dealing with large or complex data sets. Besides, as specific distributional assumptions or parametric modeling are removed, the challenge of model interpretation and prediction inference arises and has been currently at the forefront of research efforts. One problem of our interests in this regard is non-parametric or semi-parametric network estimation for data that are not independent. Specifically, influence network estimation from a multi-variate point process or time series data is a problem of fundamental importance. Prior work has focused on parametric approaches that require a known parametric model, which makes estimation procedures less robust to model mis-specification, non-linearities and heterogeneities. In Chapter 2, we develop a semi-parametric approach based on the monotone single-index multi-variate autoregressive model (SIMAM) which addresses these challenges. In particular, rather than using standard parametric approaches, we use the monotone single index model (SIM) for network estimation. We provide theoretical guarantees for dependent data, and an alternating projected gradient descent algorithm. Significantly we achieve rates of the form O(T^{-1/3} \sqrt{s\log(TM)}) (optimal in the independent design case) where s is {he number of edges in the influence network that indicates the sparsity level, M is the number of actors and T is the number of time points. In addition, we demonstrate the performance of SIMAM both on simulated data and two real data examples, and show it outperforms state-of-the-art parametric methods both in terms of prediction and network estimation. Another aspect important for distribution-free or model-free learning is the interpretation, i.e. to make the complicated non-parametric predictive models explainable. A number of model-agnostic methods for measuring variable importance (VI) have emerged in recent times, which assess the difference in predictive power between a full model trained on all variables and a reduced model that omits the variable(s) of interest. However, these methods typically encounter a bottleneck when estimating the reduced model for each variable or subset of variables, which is both costly and lacks theoretical guarantees. To address this problem, Chapter 3 proposes an efficient and adaptable approach for approximating the reduced model while ensuring important inferential guarantees. Specifically, we replace the need for fully retraining a wide neural network with a linearization that is initiated using the full model parameters. By including a ridge-like penalty to make the problem convex, we establish that our method can estimate the variable importance measure with an error rate of O({1}/{\sqrt{n}), where n represents the number of training samples, provided that the ridge penalty parameter is adequately large. Furthermore, we demonstrate that our estimator is asymptotically normal, enabling us to provide confidence bounds for the VI estimates. Finally, we demonstrate the method's speed and accuracy under different data-generating regimes and showcase its applicability in a real-world seasonal climate forecasting example. In addition to semi-parametric network estimation and fast estimation of variable importance for interpretation, an efficient method for prediction inference without specific distributional assumptions on the data is of our interest as well. In Chapter 4, we present a novel, computationally-efficient algorithm for predictive inference (PI) that requires no distributional assumptions in the data and can be computed faster than existing bootstrap-type methods for neural networks. Specifically, if there are $n$ training samples, bootstrap methods require training a model on each of the n subsamples of size n-1; for large models like neural networks, this process can be computationally prohibitive. In contrast, the proposed method trains one neural network on the full dataset with ([epsilon], [delta]) -differential privacy (DP) and then approximates each leave-one-out model efficiently using a linear approximation around the neural network estimate. With exchangeable data, we prove that our approach has a rigorous coverage guarantee that depends on the preset privacy parameters and the stability of the neural network, regardless of the data distribution. Simulations and experiments on real data demonstrate that our method satisfies the coverage guarantees with substantially reduced computation compared to bootstrap methods.




An Introduction to Causal Inference


Book Description

This paper summarizes recent advances in causal inference and underscores the paradigmatic shifts that must be undertaken in moving from traditional statistical analysis to causal analysis of multivariate data. Special emphasis is placed on the assumptions that underly all causal inferences, the languages used in formulating those assumptions, the conditional nature of all causal and counterfactual claims, and the methods that have been developed for the assessment of such claims. These advances are illustrated using a general theory of causation based on the Structural Causal Model (SCM) described in Pearl (2000a), which subsumes and unifies other approaches to causation, and provides a coherent mathematical foundation for the analysis of causes and counterfactuals. In particular, the paper surveys the development of mathematical tools for inferring (from a combination of data and assumptions) answers to three types of causal queries: (1) queries about the effects of potential interventions, (also called "causal effects" or "policy evaluation") (2) queries about probabilities of counterfactuals, (including assessment of "regret," "attribution" or "causes of effects") and (3) queries about direct and indirect effects (also known as "mediation"). Finally, the paper defines the formal and conceptual relationships between the structural and potential-outcome frameworks and presents tools for a symbiotic analysis that uses the strong features of both. The tools are demonstrated in the analyses of mediation, causes of effects, and probabilities of causation. -- p. 1.




Bayesian Networks


Book Description

Explains the material step-by-step starting from meaningful examples Steps detailed with R code in the spirit of reproducible research Real world data analyses from a Science paper reproduced and explained in detail Examples span a variety of fields across social and life sciences Overview of available software in and outside R




Proceedings of the International Workshop on Advances in Civil Aviation Systems Development


Book Description

This book includes high-quality research papers presented at International Workshop on Advances in Civil Aviation Systems Development (ACASD 2023), which was at National Aviation University, Kyiv Ukraine, on May 30, 2023. This book presents original results of a scholarly study of unique research teams and market leaders on the development in civil aviation systems and its application. The book topics include major research areas focused on advances in avionics system design, modern trends in communication, surveillance and navigation systems development, and civil avionics system maintenance questions. Also, proposed book is useful for scholars and professionals in the civil aviation domain.




Bayesian Networks


Book Description

Understand the Foundations of Bayesian Networks—Core Properties and Definitions Explained Bayesian Networks: With Examples in R introduces Bayesian networks using a hands-on approach. Simple yet meaningful examples in R illustrate each step of the modeling process. The examples start from the simplest notions and gradually increase in complexity. The authors also distinguish the probabilistic models from their estimation with data sets. The first three chapters explain the whole process of Bayesian network modeling, from structure learning to parameter learning to inference. These chapters cover discrete Bayesian, Gaussian Bayesian, and hybrid networks, including arbitrary random variables. The book then gives a concise but rigorous treatment of the fundamentals of Bayesian networks and offers an introduction to causal Bayesian networks. It also presents an overview of R and other software packages appropriate for Bayesian networks. The final chapter evaluates two real-world examples: a landmark causal protein signaling network paper and graphical modeling approaches for predicting the composition of different body parts. Suitable for graduate students and non-statisticians, this text provides an introductory overview of Bayesian networks. It gives readers a clear, practical understanding of the general approach and steps involved.




Computational and Statistical Approaches to Genomics


Book Description

The second edition of this book adds eight new contributors to reflect a modern cutting edge approach to genomics. It contains the newest research results on genomic analysis and modeling using state-of-the-art methods from engineering, statistics, and genomics. These tools and models are then applied to real biological and clinical problems. The book’s original seventeen chapters are also updated to provide new initiatives and directions.




Handbook of Markov Chain Monte Carlo


Book Description

Since their popularization in the 1990s, Markov chain Monte Carlo (MCMC) methods have revolutionized statistical computing and have had an especially profound impact on the practice of Bayesian statistics. Furthermore, MCMC methods have enabled the development and use of intricate models in an astonishing array of disciplines as diverse as fisherie




Brain Network Analysis


Book Description

This tutorial reference serves as a coherent overview of various statistical and mathematical approaches used in brain network analysis, where modeling the complex structures and functions of the human brain often poses many unique computational and statistical challenges. This book fills a gap as a textbook for graduate students while simultaneously articulating important and technically challenging topics. Whereas most available books are graph theory-centric, this text introduces techniques arising from graph theory and expands to include other different models in its discussion on network science, regression, and algebraic topology. Links are included to the sample data and codes used in generating the book's results and figures, helping to empower methodological understanding in a manner immediately usable to both researchers and students.