Predictive Inference


Book Description

The author's research has been directed towards inference involving observables rather than parameters. In this book, he brings together his views on predictive or observable inference and its advantages over parametric inference. While the book discusses a variety of approaches to prediction including those based on parametric, nonparametric, and nonstochastic statistical models, it is devoted mainly to predictive applications of the Bayesian approach. It not only substitutes predictive analyses for parametric analyses, but it also presents predictive analyses that have no real parametric analogues. It demonstrates that predictive inference can be a critical component of even strict parametric inference when dealing with interim analyses. This approach to predictive inference will be of interest to statisticians, psychologists, econometricians, and sociologists.




Predictive Statistics


Book Description

A bold retooling of statistics to focus directly on predictive performance with traditional and contemporary data types and methodologies.




The Elements of Statistical Learning


Book Description

During the past decade there has been an explosion in computation and information technology. With it have come vast amounts of data in a variety of fields such as medicine, biology, finance, and marketing. The challenge of understanding these data has led to the development of new tools in the field of statistics, and spawned new areas such as data mining, machine learning, and bioinformatics. Many of these tools have common underpinnings but are often expressed with different terminology. This book describes the important ideas in these areas in a common conceptual framework. While the approach is statistical, the emphasis is on concepts rather than mathematics. Many examples are given, with a liberal use of color graphics. It should be a valuable resource for statisticians and anyone interested in data mining in science or industry. The book’s coverage is broad, from supervised learning (prediction) to unsupervised learning. The many topics include neural networks, support vector machines, classification trees and boosting---the first comprehensive treatment of this topic in any book. This major new edition features many topics not covered in the original, including graphical models, random forests, ensemble methods, least angle regression & path algorithms for the lasso, non-negative matrix factorization, and spectral clustering. There is also a chapter on methods for “wide” data (p bigger than n), including multiple testing and false discovery rates. Trevor Hastie, Robert Tibshirani, and Jerome Friedman are professors of statistics at Stanford University. They are prominent researchers in this area: Hastie and Tibshirani developed generalized additive models and wrote a popular book of that title. Hastie co-developed much of the statistical modeling software and environment in R/S-PLUS and invented principal curves and surfaces. Tibshirani proposed the lasso and is co-author of the very successful An Introduction to the Bootstrap. Friedman is the co-inventor of many data-mining tools including CART, MARS, projection pursuit and gradient boosting.




Probably Not


Book Description

An engaging, entertaining, and informative introduction to probability and prediction in our everyday lives Although Probably Not deals with probability and statistics, it is not heavily mathematical and is not filled with complex derivations, proofs, and theoretical problem sets. This book unveils the world of statistics through questions such as what is known based upon the information at hand and what can be expected to happen. While learning essential concepts including "the confidence factor" and "random walks," readers will be entertained and intrigued as they move from chapter to chapter. Moreover, the author provides a foundation of basic principles to guide decision making in almost all facets of life including playing games, developing winning business strategies, and managing personal finances. Much of the book is organized around easy-to-follow examples that address common, everyday issues such as: How travel time is affected by congestion, driving speed, and traffic lights Why different gambling casino strategies ultimately offer players no advantage How to estimate how many different birds of one species are seen on a walk through the woods Seemingly random events—coin flip games, the Central Limit Theorem, binomial distributions and Poisson distributions, Parrando's Paradox, and Benford's Law—are addressed and treated through key concepts and methods in probability. In addition, fun-to-solve problems including "the shared birthday" and "the prize behind door number one, two, or three" are found throughout the book, which allow readers to test and practice their new probability skills. Requiring little background knowledge of mathematics, readers will gain a greater understanding of the many daily activities and events that involve random processes and statistics. Combining the mathematics of probability with real-world examples, Probably Not is an ideal reference for practitioners and students who would like to learn more about the role of probability and statistics in everyday decision making.




Large-Scale Inference


Book Description

We live in a new age for statistical inference, where modern scientific technology such as microarrays and fMRI machines routinely produce thousands and sometimes millions of parallel data sets, each with its own estimation or testing problem. Doing thousands of problems at once is more than repeated application of classical methods. Taking an empirical Bayes approach, Bradley Efron, inventor of the bootstrap, shows how information accrues across problems in a way that combines Bayesian and frequentist ideas. Estimation, testing and prediction blend in this framework, producing opportunities for new methodologies of increased power. New difficulties also arise, easily leading to flawed inferences. This book takes a careful look at both the promise and pitfalls of large-scale statistical inference, with particular attention to false discovery rates, the most successful of the new statistical techniques. Emphasis is on the inferential ideas underlying technical developments, illustrated using a large number of real examples.




Modern Statistics with R


Book Description

The past decades have transformed the world of statistical data analysis, with new methods, new types of data, and new computational tools. Modern Statistics with R introduces you to key parts of this modern statistical toolkit. It teaches you: Data wrangling - importing, formatting, reshaping, merging, and filtering data in R. Exploratory data analysis - using visualisations and multivariate techniques to explore datasets. Statistical inference - modern methods for testing hypotheses and computing confidence intervals. Predictive modelling - regression models and machine learning methods for prediction, classification, and forecasting. Simulation - using simulation techniques for sample size computations and evaluations of statistical methods. Ethics in statistics - ethical issues and good statistical practice. R programming - writing code that is fast, readable, and (hopefully!) free from bugs. No prior programming experience is necessary. Clear explanations and examples are provided to accommodate readers at all levels of familiarity with statistical principles and coding practices. A basic understanding of probability theory can enhance comprehension of certain concepts discussed within this book. In addition to plenty of examples, the book includes more than 200 exercises, with fully worked solutions available at: www.modernstatisticswithr.com.




Inferences during Reading


Book Description

Inferencing is defined as 'the act of deriving logical conclusions from premises known or assumed to be true', and it is one of the most important processes necessary for successful comprehension during reading. This volume features contributions by distinguished researchers in cognitive psychology, educational psychology, and neuroscience on topics central to our understanding of the inferential process during reading. The chapters cover aspects of inferencing that range from the fundamental bottom-up processes that form the basis for an inference to occur, to the more strategic processes that transpire when a reader is engaged in literary understanding of a text. Basic activation mechanisms, word-level inferencing, methodological considerations, inference validation, causal inferencing, emotion, development of inferences processes as a skill, embodiment, contributions from neuroscience, and applications to naturalistic text are all covered as well as expository text, online learning materials, and literary immersion.




Applied Predictive Modeling


Book Description

Applied Predictive Modeling covers the overall predictive modeling process, beginning with the crucial steps of data preprocessing, data splitting and foundations of model tuning. The text then provides intuitive explanations of numerous common and modern regression and classification techniques, always with an emphasis on illustrating and solving real data problems. The text illustrates all parts of the modeling process through many hands-on, real-life examples, and every chapter contains extensive R code for each step of the process. This multi-purpose text can be used as an introduction to predictive models and the overall modeling process, a practitioner’s reference handbook, or as a text for advanced undergraduate or graduate level predictive modeling courses. To that end, each chapter contains problem sets to help solidify the covered concepts and uses data available in the book’s R package. This text is intended for a broad audience as both an introduction to predictive models as well as a guide to applying them. Non-mathematical readers will appreciate the intuitive explanations of the techniques while an emphasis on problem-solving with real data across a wide variety of applications will aid practitioners who wish to extend their expertise. Readers should have knowledge of basic statistical ideas, such as correlation and linear regression analysis. While the text is biased against complex equations, a mathematical background is needed for advanced topics.




Bayesian Methods in Finance


Book Description

Bayesian Methods in Finance provides a detailed overview of the theory of Bayesian methods and explains their real-world applications to financial modeling. While the principles and concepts explained throughout the book can be used in financial modeling and decision making in general, the authors focus on portfolio management and market risk management—since these are the areas in finance where Bayesian methods have had the greatest penetration to date.




Situation Models and Levels of Coherence


Book Description

"Situation Models and Levels of Coherence proposes an integrated view of the various theoretical approaches to discourse comprehension, and in particular of situation-model building, as evidenced by empirical findings and computational models. The objective has been to arrive at an in-depth definition of the internal structure of situation models, and of the cognitive processes that underlie their elaboration, while articulating these two aspects of the reader's mental representations by bringing to the fore the core concept of coherence." (cit. 4e p. de couv.).