Identification for Prediction and Decision


Book Description

This book is a full-scale exposition of Charles Manski's new methodology for analyzing empirical questions in the social sciences. He recommends that researchers first ask what can be learned from data alone, and then ask what can be learned when data are combined with credible weak assumptions. Inferences predicated on weak assumptions, he argues, can achieve wide consensus, while ones that require strong assumptions almost inevitably are subject to sharp disagreements. Building on the foundation laid in the author's Identification Problems in the Social Sciences (Harvard, 1995), the book's fifteen chapters are organized in three parts. Part I studies prediction with missing or otherwise incomplete data. Part II concerns the analysis of treatment response, which aims to predict outcomes when alternative treatment rules are applied to a population. Part III studies prediction of choice behavior. Each chapter juxtaposes developments of methodology with empirical or numerical illustrations. The book employs a simple notation and mathematical apparatus, using only basic elements of probability theory.




Public Policy in an Uncertain World


Book Description

Manski argues that public policy is based on untrustworthy analysis. Failing to account for uncertainty in an uncertain world, policy analysis routinely misleads policy makers with expressions of certitude. Manski critiques the status quo and offers an innovation to improve both how policy research is conducted and how it is used by policy makers.




A Course in Econometrics


Book Description

This text prepares first-year graduate students and advanced undergraduates for empirical research in economics, and also equips them for specialization in econometric theory, business, and sociology. A Course in Econometrics is likely to be the text most thoroughly attuned to the needs of your students. Derived from the course taught by Arthur S. Goldberger at the University of Wisconsin-Madison and at Stanford University, it is specifically designed for use over two semesters, offers students the most thorough grounding in introductory statistical inference, and offers a substantial amount of interpretive material. The text brims with insights, strikes a balance between rigor and intuition, and provokes students to form their own critical opinions. A Course in Econometrics thoroughly covers the fundamentals--classical regression and simultaneous equations--and offers clear and logical explorations of asymptotic theory and nonlinear regression. To accommodate students with various levels of preparation, the text opens with a thorough review of statistical concepts and methods, then proceeds to the regression model and its variants. Bold subheadings introduce and highlight key concepts throughout each chapter. Each chapter concludes with a set of exercises specifically designed to reinforce and extend the material covered. Many of the exercises include real microdata analyses, and all are ideally suited to use as homework and test questions.




Identification Problems in the Social Sciences


Book Description

The author draws on examples from a range of disciplines to provide social and behavioural scientists with a toolkit for finding bounds when predicting behaviours based upon nonexperimental and experimental data.




Patient Care Under Uncertainty


Book Description

For the past few years, the author, a renowned economist, has been applying the statistical tools of economics to decision making under uncertainty in the context of patient health status and response to treatment. He shows how statistical imprecision and identification problems affect empirical research in the patient-care sphere.




Discrete Choice Methods with Simulation


Book Description

This book describes the new generation of discrete choice methods, focusing on the many advances that are made possible by simulation. Researchers use these statistical methods to examine the choices that consumers, households, firms, and other agents make. Each of the major models is covered: logit, generalized extreme value, or GEV (including nested and cross-nested logits), probit, and mixed logit, plus a variety of specifications that build on these basics. Simulation-assisted estimation procedures are investigated and compared, including maximum stimulated likelihood, method of simulated moments, and method of simulated scores. Procedures for drawing from densities are described, including variance reduction techniques such as anithetics and Halton draws. Recent advances in Bayesian procedures are explored, including the use of the Metropolis-Hastings algorithm and its variant Gibbs sampling. The second edition adds chapters on endogeneity and expectation-maximization (EM) algorithms. No other book incorporates all these fields, which have arisen in the past 25 years. The procedures are applicable in many fields, including energy, transportation, environmental studies, health, labor, and marketing.




Interpretable Machine Learning


Book Description

This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.




The Elements of Statistical Learning


Book Description

During the past decade there has been an explosion in computation and information technology. With it have come vast amounts of data in a variety of fields such as medicine, biology, finance, and marketing. The challenge of understanding these data has led to the development of new tools in the field of statistics, and spawned new areas such as data mining, machine learning, and bioinformatics. Many of these tools have common underpinnings but are often expressed with different terminology. This book describes the important ideas in these areas in a common conceptual framework. While the approach is statistical, the emphasis is on concepts rather than mathematics. Many examples are given, with a liberal use of color graphics. It should be a valuable resource for statisticians and anyone interested in data mining in science or industry. The book’s coverage is broad, from supervised learning (prediction) to unsupervised learning. The many topics include neural networks, support vector machines, classification trees and boosting---the first comprehensive treatment of this topic in any book. This major new edition features many topics not covered in the original, including graphical models, random forests, ensemble methods, least angle regression & path algorithms for the lasso, non-negative matrix factorization, and spectral clustering. There is also a chapter on methods for “wide” data (p bigger than n), including multiple testing and false discovery rates. Trevor Hastie, Robert Tibshirani, and Jerome Friedman are professors of statistics at Stanford University. They are prominent researchers in this area: Hastie and Tibshirani developed generalized additive models and wrote a popular book of that title. Hastie co-developed much of the statistical modeling software and environment in R/S-PLUS and invented principal curves and surfaces. Tibshirani proposed the lasso and is co-author of the very successful An Introduction to the Bootstrap. Friedman is the co-inventor of many data-mining tools including CART, MARS, projection pursuit and gradient boosting.




Surfing Uncertainty


Book Description

Exciting new theories in neuroscience, psychology, and artificial intelligence are revealing minds like ours as predictive minds, forever trying to guess the incoming streams of sensory stimulation before they arrive. In this up-to-the-minute treatment, philosopher and cognitive scientist Andy Clark explores new ways of thinking about perception, action, and the embodied mind.




Neural Networks for Identification, Prediction and Control


Book Description

In recent years, there has been a growing interest in applying neural networks to dynamic systems identification (modelling), prediction and control. Neural networks are computing systems characterised by the ability to learn from examples rather than having to be programmed in a conventional sense. Their use enables the behaviour of complex systems to be modelled and predicted and accurate control to be achieved through training, without a priori information about the systems' structures or parameters. This book describes examples of applications of neural networks In modelling, prediction and control. The topics covered include identification of general linear and non-linear processes, forecasting of river levels, stock market prices and currency exchange rates, and control of a time-delayed plant and a two-joint robot. These applications employ the major types of neural networks and learning algorithms. The neural network types considered in detail are the muhilayer perceptron (MLP), the Elman and Jordan networks and the Group-Method-of-Data-Handling (GMDH) network. In addition, cerebellar-model-articulation-controller (CMAC) networks and neuromorphic fuzzy logic systems are also presented. The main learning algorithm adopted in the applications is the standard backpropagation (BP) algorithm. Widrow-Hoff learning, dynamic BP and evolutionary learning are also described.