Data quality assurance. Module 3. Site assessment of data quality


Book Description

This publication is one of the three module toolkit and provide technical guidance and tools to support the work on strengthening data quality in countries. This is part of the Division of Data, Analytics and Delivery for Impact’s scope of work providing normative guidance for health information system strengthening.




Data Quality


Book Description

Poor data quality can seriously hinder or damage the efficiency and effectiveness of organizations and businesses. The growing awareness of such repercussions has led to major public initiatives like the "Data Quality Act" in the USA and the "European 2003/98" directive of the European Parliament. Batini and Scannapieco present a comprehensive and systematic introduction to the wide set of issues related to data quality. They start with a detailed description of different data quality dimensions, like accuracy, completeness, and consistency, and their importance in different types of data, like federated data, web data, or time-dependent data, and in different data categories classified according to frequency of change, like stable, long-term, and frequently changing data. The book's extensive description of techniques and methodologies from core data quality research as well as from related fields like data mining, probability theory, statistical data analysis, and machine learning gives an excellent overview of the current state of the art. The presentation is completed by a short description and critical comparison of tools and practical methodologies, which will help readers to resolve their own quality problems. This book is an ideal combination of the soundness of theoretical foundations and the applicability of practical approaches. It is ideally suited for everyone – researchers, students, or professionals – interested in a comprehensive overview of data quality issues. In addition, it will serve as the basis for an introductory course or for self-study on this topic.




The Practitioner's Guide to Data Quality Improvement


Book Description

The Practitioner's Guide to Data Quality Improvement offers a comprehensive look at data quality for business and IT, encompassing people, process, and technology. It shares the fundamentals for understanding the impacts of poor data quality, and guides practitioners and managers alike in socializing, gaining sponsorship for, planning, and establishing a data quality program. It demonstrates how to institute and run a data quality program, from first thoughts and justifications to maintenance and ongoing metrics. It includes an in-depth look at the use of data quality tools, including business case templates, and tools for analysis, reporting, and strategic planning. This book is recommended for data management practitioners, including database analysts, information analysts, data administrators, data architects, enterprise architects, data warehouse engineers, and systems analysts, and their managers. - Offers a comprehensive look at data quality for business and IT, encompassing people, process, and technology. - Shows how to institute and run a data quality program, from first thoughts and justifications to maintenance and ongoing metrics. - Includes an in-depth look at the use of data quality tools, including business case templates, and tools for analysis, reporting, and strategic planning.







Data quality assurance. Module 2. Discrete desk review of data quality


Book Description

This publication is one of the three module toolkit and provide technical guidance and tools to support the work on strengthening data quality in countries. This is part of the Division of Data and Delivery for Impact's scope of work providing normative guidance for health information system strengthening.




Handbook of Data Quality


Book Description

The issue of data quality is as old as data itself. However, the proliferation of diverse, large-scale and often publically available data on the Web has increased the risk of poor data quality and misleading data interpretations. On the other hand, data is now exposed at a much more strategic level e.g. through business intelligence systems, increasing manifold the stakes involved for individuals, corporations as well as government agencies. There, the lack of knowledge about data accuracy, currency or completeness can have erroneous and even catastrophic results. With these changes, traditional approaches to data management in general, and data quality control specifically, are challenged. There is an evident need to incorporate data quality considerations into the whole data cycle, encompassing managerial/governance as well as technical aspects. Data quality experts from research and industry agree that a unified framework for data quality management should bring together organizational, architectural and computational approaches. Accordingly, Sadiq structured this handbook in four parts: Part I is on organizational solutions, i.e. the development of data quality objectives for the organization, and the development of strategies to establish roles, processes, policies, and standards required to manage and ensure data quality. Part II, on architectural solutions, covers the technology landscape required to deploy developed data quality management processes, standards and policies. Part III, on computational solutions, presents effective and efficient tools and techniques related to record linkage, lineage and provenance, data uncertainty, and advanced integrity constraints. Finally, Part IV is devoted to case studies of successful data quality initiatives that highlight the various aspects of data quality in action. The individual chapters present both an overview of the respective topic in terms of historical research and/or practice and state of the art, as well as specific techniques, methodologies and frameworks developed by the individual contributors. Researchers and students of computer science, information systems, or business management as well as data professionals and practitioners will benefit most from this handbook by not only focusing on the various sections relevant to their research area or particular practical work, but by also studying chapters that they may initially consider not to be directly relevant to them, as there they will learn about new perspectives and approaches.







Software Process Improvement and Capability Determination


Book Description

This book constitutes the refereed proceedings of the 16th International Conference on Software Process Improvement and Capability Determination, SPICE 2016, held in Dublin, Ireland, in June 2016. The 28 full papers presented together with 5 short papers were carefully reviewed and selected from 52 submissions. The papers are organized in the following topical sections: SPI in regulated and safety critical domains; gamification and education issues in SPI; SPI in agile and small settings; SPI and assessment; SPI and project management concerns; empirical research case studies of SPI; knowledge and human communications issues in SPI.




Assessing the quality of agricultural market information systems: A self-assessment guide


Book Description

Over approximately the past 40 years, many developing countries invested in the establishment of agricultural market information systems or services (MIS). These systems or services were initially run by government agencies, but since the turn of the millennium private organizations have shown interest in providing data on a commercial basis. To date, however, these private services, while usually being more efficient than the government-run ones, have also largely depended on donor support for their continued operation. It has proved difficult to develop a profitable business model as many of the clients are small farmers and traders. Agricultural market information systems or services (MIS) can cover staples, horticultural crops, livestock, and export commodities. They are generally designed to collect, process, and disseminate or distribute data of relevance to farmers, traders and other buyers, such as processors, but the data they generate can also be used for a variety of purposes by governments, donors, international organizations and others.