Nowcasting GDP - A Scalable Approach Using DFM, Machine Learning and Novel Data, Applied to European Economies


Book Description

This paper describes recent work to strengthen nowcasting capacity at the IMF’s European department. It motivates and compiles datasets of standard and nontraditional variables, such as Google search and air quality. It applies standard dynamic factor models (DFMs) and several machine learning (ML) algorithms to nowcast GDP growth across a heterogenous group of European economies during normal and crisis times. Most of our methods significantly outperform the AR(1) benchmark model. Our DFMs tend to perform better during normal times while many of the ML methods we used performed strongly at identifying turning points. Our approach is easily applicable to other countries, subject to data availability.




Forecasting National Activity Using Lots of International Predictors


Book Description

We look at how large international datasets can improve forecasts of national activity. We use the case of New Zealand, an archetypal small open economy. We apply "data-rich" factor and shrinkage methods to tackle the problem of efficiently handling hundreds of predictor data series from many countries. The methods covered are principal components, targeted predictors, weighted principal components, partial least squares, elastic net and ridge regression. Using these methods, we assess the marginal predictive content of international data for New Zealand GDP growth. We find that exploiting a large number of international predictors can improve forecasts of our target variable, compared to more traditional models based on small datasets. This is in spite of New Zealand survey data capturing a substantial proportion of the predictive information in the international data. The largest forecasting accuracy gains from including international predictors are at longer forecast horizons. The forecasting performance achievable with the data-rich methods differs widely, with shrinkage methods and partial least squares performing best. We also assess the type of international data that contains the most predictive information for New Zealand growth over our sample.




Data Science for Economics and Finance


Book Description

This open access book covers the use of data science, including advanced machine learning, big data analytics, Semantic Web technologies, natural language processing, social media analysis, time series analysis, among others, for applications in economics and finance. In addition, it shows some successful applications of advanced data science solutions used to extract new knowledge from data in order to improve economic forecasting models. The book starts with an introduction on the use of data science technologies in economics and finance and is followed by thirteen chapters showing success stories of the application of specific data science methodologies, touching on particular topics related to novel big data sources and technologies for economic analysis (e.g. social media and news); big data models leveraging on supervised/unsupervised (deep) machine learning; natural language processing to build economic and financial indicators; and forecasting and nowcasting of economic variables through time series analysis. This book is relevant to all stakeholders involved in digital and data-intensive research in economics and finance, helping them to understand the main opportunities and challenges, become familiar with the latest methodological findings, and learn how to use and evaluate the performances of novel tools and frameworks. It primarily targets data scientists and business analysts exploiting data science technologies, and it will also be a useful resource to research students in disciplines and courses related to these topics. Overall, readers will learn modern and effective data science solutions to create tangible innovations for economic and financial applications.




Powering the Digital Economy: Opportunities and Risks of Artificial Intelligence in Finance


Book Description

This paper discusses the impact of the rapid adoption of artificial intelligence (AI) and machine learning (ML) in the financial sector. It highlights the benefits these technologies bring in terms of financial deepening and efficiency, while raising concerns about its potential in widening the digital divide between advanced and developing economies. The paper advances the discussion on the impact of this technology by distilling and categorizing the unique risks that it could pose to the integrity and stability of the financial system, policy challenges, and potential regulatory approaches. The evolving nature of this technology and its application in finance means that the full extent of its strengths and weaknesses is yet to be fully understood. Given the risk of unexpected pitfalls, countries will need to strengthen prudential oversight.




Macroeconomic Forecasting in the Era of Big Data


Book Description

This book surveys big data tools used in macroeconomic forecasting and addresses related econometric issues, including how to capture dynamic relationships among variables; how to select parsimonious models; how to deal with model uncertainty, instability, non-stationarity, and mixed frequency data; and how to evaluate forecasts, among others. Each chapter is self-contained with references, and provides solid background information, while also reviewing the latest advances in the field. Accordingly, the book offers a valuable resource for researchers, professional forecasters, and students of quantitative economics.




The Oxford Handbook of Economic Forecasting


Book Description

Greater data availability has been coupled with developments in statistical theory and economic theory to allow more elaborate and complicated models to be entertained. These include factor models, DSGE models, restricted vector autoregressions, and non-linear models.




U-MIDAS


Book Description




Genetic Programming Theory and Practice X


Book Description

These contributions, written by the foremost international researchers and practitioners of Genetic Programming (GP), explore the synergy between theoretical and empirical results on real-world problems, producing a comprehensive view of the state of the art in GP. Topics in this volume include: evolutionary constraints, relaxation of selection mechanisms, diversity preservation strategies, flexing fitness evaluation, evolution in dynamic environments, multi-objective and multi-modal selection, foundations of evolvability, evolvable and adaptive evolutionary operators, foundation of injecting expert knowledge in evolutionary search, analysis of problem difficulty and required GP algorithm complexity, foundations in running GP on the cloud – communication, cooperation, flexible implementation, and ensemble methods. Additional focal points for GP symbolic regression are: (1) The need to guarantee convergence to solutions in the function discovery mode; (2) Issues on model validation; (3) The need for model analysis workflows for insight generation based on generated GP solutions – model exploration, visualization, variable selection, dimensionality analysis; (4) Issues in combining different types of data. Readers will discover large-scale, real-world applications of GP to a variety of problem domains via in-depth presentations of the latest and most significant results.




Big Data for Twenty-First-Century Economic Statistics


Book Description

Introduction.Big data for twenty-first-century economic statistics: the future is now /Katharine G. Abraham, Ron S. Jarmin, Brian C. Moyer, and Matthew D. Shapiro --Toward comprehensive use of big data in economic statistics.Reengineering key national economic indicators /Gabriel Ehrlich, John Haltiwanger, Ron S. Jarmin, David Johnson, and Matthew D. Shapiro ;Big data in the US consumer price index: experiences and plans /Crystal G. Konny, Brendan K. Williams, and David M. Friedman ;Improving retail trade data products using alternative data sources /Rebecca J. Hutchinson ;From transaction data to economic statistics: constructing real-time, high-frequency, geographic measures of consumer spending /Aditya Aladangady, Shifrah Aron-Dine, Wendy Dunn, Laura Feiveson, Paul Lengermann, and Claudia Sahm ;Improving the accuracy of economic measurement with multiple data sources: the case of payroll employment data /Tomaz Cajner, Leland D. Crane, Ryan A. Decker, Adrian Hamins-Puertolas, and Christopher Kurz --Uses of big data for classification.Transforming naturally occurring text data into economic statistics: the case of online job vacancy postings /Arthur Turrell, Bradley Speigner, Jyldyz Djumalieva, David Copple, and James Thurgood ;Automating response evaluation for franchising questions on the 2017 economic census /Joseph Staudt, Yifang Wei, Lisa Singh, Shawn Klimek, J. Bradford Jensen, and Andrew Baer ;Using public data to generate industrial classification codes /John Cuffe, Sudip Bhattacharjee, Ugochukwu Etudo, Justin C. Smith, Nevada Basdeo, Nathaniel Burbank, and Shawn R. Roberts --Uses of big data for sectoral measurement.Nowcasting the local economy: using Yelp data to measure economic activity /Edward L. Glaeser, Hyunjin Kim, and Michael Luca ;Unit values for import and export price indexes: a proof of concept /Don A. Fast and Susan E. Fleck ;Quantifying productivity growth in the delivery of important episodes of care within the Medicare program using insurance claims and administrative data /John A. Romley, Abe Dunn, Dana Goldman, and Neeraj Sood ;Valuing housing services in the era of big data: a user cost approach leveraging Zillow microdata /Marina Gindelsky, Jeremy G. Moulton, and Scott A. Wentland --Methodological challenges and advances.Off to the races: a comparison of machine learning and alternative data for predicting economic indicators /Jeffrey C. Chen, Abe Dunn, Kyle Hood, Alexander Driessen, and Andrea Batch ;A machine learning analysis of seasonal and cyclical sales in weekly scanner data /Rishab Guha and Serena Ng ;Estimating the benefits of new products /W. Erwin Diewert and Robert C. Feenstra.




Completing the Market: Generating Shadow CDS Spreads by Machine Learning


Book Description

We compared the predictive performance of a series of machine learning and traditional methods for monthly CDS spreads, using firms’ accounting-based, market-based and macroeconomics variables for a time period of 2006 to 2016. We find that ensemble machine learning methods (Bagging, Gradient Boosting and Random Forest) strongly outperform other estimators, and Bagging particularly stands out in terms of accuracy. Traditional credit risk models using OLS techniques have the lowest out-of-sample prediction accuracy. The results suggest that the non-linear machine learning methods, especially the ensemble methods, add considerable value to existent credit risk prediction accuracy and enable CDS shadow pricing for companies missing those securities.