Data Governance


Book Description

Managing data continues to grow as a necessity for modern organizations. There are seemingly infinite opportunities for organic growth, reduction of costs, and creation of new products and services. It has become apparent that none of these opportunities can happen smoothly without data governance. The cost of exponential data growth and privacy / security concerns are becoming burdensome. Organizations will encounter unexpected consequences in new sources of risk. The solution to these challenges is also data governance; ensuring balance between risk and opportunity. Data Governance, Second Edition, is for any executive, manager or data professional who needs to understand or implement a data governance program. It is required to ensure consistent, accurate and reliable data across their organization. This book offers an overview of why data governance is needed, how to design, initiate, and execute a program and how to keep the program sustainable. This valuable resource provides comprehensive guidance to beginning professionals, managers or analysts looking to improve their processes, and advanced students in Data Management and related courses. With the provided framework and case studies all professionals in the data governance field will gain key insights into launching successful and money-saving data governance program. - Incorporates industry changes, lessons learned and new approaches - Explores various ways in which data analysts and managers can ensure consistent, accurate and reliable data across their organizations - Includes new case studies which detail real-world situations - Explores all of the capabilities an organization must adopt to become data driven - Provides guidance on various approaches to data governance, to determine whether an organization should be low profile, central controlled, agile, or traditional - Provides guidance on using technology and separating vendor hype from sincere delivery of necessary capabilities - Offers readers insights into how their organizations can improve the value of their data, through data quality, data strategy and data literacy - Provides up to 75% brand-new content compared to the first edition




Data Governance


Book Description




Managing Data Quality


Book Description

This book explains data quality management in practical terms, focusing on three key areas - the nature of data in enterprises, the purpose and scope of data quality management, and implementing a data quality management system, in line with ISO 8000-61. Examples of good practice in data quality management are also included.




Registries for Evaluating Patient Outcomes


Book Description

This User’s Guide is intended to support the design, implementation, analysis, interpretation, and quality evaluation of registries created to increase understanding of patient outcomes. For the purposes of this guide, a patient registry is an organized system that uses observational study methods to collect uniform data (clinical and other) to evaluate specified outcomes for a population defined by a particular disease, condition, or exposure, and that serves one or more predetermined scientific, clinical, or policy purposes. A registry database is a file (or files) derived from the registry. Although registries can serve many purposes, this guide focuses on registries created for one or more of the following purposes: to describe the natural history of disease, to determine clinical effectiveness or cost-effectiveness of health care products and services, to measure or monitor safety and harm, and/or to measure quality of care. Registries are classified according to how their populations are defined. For example, product registries include patients who have been exposed to biopharmaceutical products or medical devices. Health services registries consist of patients who have had a common procedure, clinical encounter, or hospitalization. Disease or condition registries are defined by patients having the same diagnosis, such as cystic fibrosis or heart failure. The User’s Guide was created by researchers affiliated with AHRQ’s Effective Health Care Program, particularly those who participated in AHRQ’s DEcIDE (Developing Evidence to Inform Decisions About Effectiveness) program. Chapters were subject to multiple internal and external independent reviews.




Executing Data Quality Projects


Book Description

Executing Data Quality Projects, Second Edition presents a structured yet flexible approach for creating, improving, sustaining and managing the quality of data and information within any organization. Studies show that data quality problems are costing businesses billions of dollars each year, with poor data linked to waste and inefficiency, damaged credibility among customers and suppliers, and an organizational inability to make sound decisions. Help is here! This book describes a proven Ten Step approach that combines a conceptual framework for understanding information quality with techniques, tools, and instructions for practically putting the approach to work – with the end result of high-quality trusted data and information, so critical to today's data-dependent organizations. The Ten Steps approach applies to all types of data and all types of organizations – for-profit in any industry, non-profit, government, education, healthcare, science, research, and medicine. This book includes numerous templates, detailed examples, and practical advice for executing every step. At the same time, readers are advised on how to select relevant steps and apply them in different ways to best address the many situations they will face. The layout allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, best practices, and warnings. The experience of actual clients and users of the Ten Steps provide real examples of outputs for the steps plus highlighted, sidebar case studies called Ten Steps in Action. This book uses projects as the vehicle for data quality work and the word broadly to include: 1) focused data quality improvement projects, such as improving data used in supply chain management, 2) data quality activities in other projects such as building new applications and migrating data from legacy systems, integrating data because of mergers and acquisitions, or untangling data due to organizational breakups, and 3) ad hoc use of data quality steps, techniques, or activities in the course of daily work. The Ten Steps approach can also be used to enrich an organization's standard SDLC (whether sequential or Agile) and it complements general improvement methodologies such as six sigma or lean. No two data quality projects are the same but the flexible nature of the Ten Steps means the methodology can be applied to all. The new Second Edition highlights topics such as artificial intelligence and machine learning, Internet of Things, security and privacy, analytics, legal and regulatory requirements, data science, big data, data lakes, and cloud computing, among others, to show their dependence on data and information and why data quality is more relevant and critical now than ever before. - Includes concrete instructions, numerous templates, and practical advice for executing every step of The Ten Steps approach - Contains real examples from around the world, gleaned from the author's consulting practice and from those who implemented based on her training courses and the earlier edition of the book - Allows for quick reference with an easy-to-use format highlighting key concepts and definitions, important checkpoints, communication activities, and best practices - A companion Web site includes links to numerous data quality resources, including many of the templates featured in the text, quick summaries of key ideas from the Ten Steps methodology, and other tools and information that are available online




Module for assessing and strengthening the quality of viral load testing data within HIV programmes and patient monitoring systems: implementation tool, Second edition


Book Description

This technical update of the 2020 viral load data quality module provides further guidance on the recommended data quality assurance activities, updated web annexes to support country implementation, and generic budgets for viral load testing data quality monitoring activities. These are part of ongoing efforts to standardize approaches to ensure that accurate and timely HIV viral load testing data and results are available for both clinical use and to strengthen programme monitoring. This aligns with recommendations outlined in the 2022 WHO Consolidated guidelines on person-centred HIV strategic information on data quality assessment and improvement. The updated guidance on the priority indicators and the key elements of data quality include: - Key indicators to be included in data quality monitoring activities; - Main activities to be included in the viral load testing data quality assessment process; - The calculation of viral load test turnaround times; - Considerations for data quality assessments for sites with electronic data systems; - Sampling for national data quality assessments of sites and clinical records; - Data quality monitoring via lot-quality assurance sampling; - Considerations for facilities with point-of-care or near point-of-care viral load testing; - Considerations for data quality assessments of viral load testing data for pregnant and breastfeeding women; and - Recording the limitations and challenges of data quality monitoring assurance activities.




Data Management in Large-Scale Education Research


Book Description

Research data management is becoming more complicated. Researchers are collecting more data, using more complex technologies, all the while increasing the visibility of our work with the push for data sharing and open science practices. Ad hoc data management practices may have worked for us in the past, but now others need to understand our processes as well, requiring researchers to be more thoughtful in planning their data management routines. This book is for anyone involved in a research study involving original data collection. While the book focuses on quantitative data, typically collected from human participants, many of the practices covered can apply to other types of data as well. The book contains foundational context, instructions, and practical examples to help researchers in the field of education begin to understand how to create data management workflows for large-scale, typically federally funded, research studies. The book starts by describing the research life cycle and how data management fits within this larger picture. The remaining chapters are then organized by each phase of the life cycle, with examples of best practices provided for each phase. Finally, considerations on whether the reader should implement, and how to integrate those practices into a workflow, are discussed. Key Features: Provides a holistic approach to the research life cycle, showing how project management and data management processes work in parallel and collaboratively Can be read in its entirety, or referenced as needed throughout the life cycle Includes relatable examples specific to education research Includes a discussion on how to organize and document data in preparation for data sharing requirements Contains links to example documents as well as templates to help readers implement practices




Driving Data Quality with Data Contracts


Book Description

Everything you need to know to apply data contracts and build a truly data-driven organization that harnesses quality data to deliver tangible business value Purchase of the print or Kindle book includes a free PDF eBook Key Features Understand data contracts and their power to resolving the problems in contemporary data platforms Learn how to design and implement a cutting-edge data platform powered by data contracts Access practical guidance from the pioneer of data contracts to get expert insights on effective utilization Book DescriptionDespite the passage of time and the evolution of technology and architecture, the challenges we face in building data platforms persist. Our data often remains unreliable, lacks trust, and fails to deliver the promised value. With Driving Data Quality with Data Contracts, you’ll discover the potential of data contracts to transform how you build your data platforms, finally overcoming these enduring problems. You’ll learn how establishing contracts as the interface allows you to explicitly assign responsibility and accountability of the data to those who know it best—the data generators—and give them the autonomy to generate and manage data as required. The book will show you how data contracts ensure that consumers get quality data with clearly defined expectations, enabling them to build on that data with confidence to deliver valuable analytics, performant ML models, and trusted data-driven products. By the end of this book, you’ll have gained a comprehensive understanding of how data contracts can revolutionize your organization’s data culture and provide a competitive advantage by unlocking the real value within your data.What you will learn Gain insights into the intricacies and shortcomings of today's data architectures Understand exactly how data contracts can solve prevalent data challenges Drive a fundamental transformation of your data culture by implementing data contracts Discover what goes into a data contract and why it's important Design a modern data architecture that leverages the power of data contracts Explore sample implementations to get practical knowledge of using data contracts Embrace best practices for the successful deployment of data contracts Who this book is for If you’re a data engineer, data leader, architect, or practitioner thinking about your data architecture and looking to design one that enables your organization to get the most value from your data, this book is for you. Additionally, staff engineers, product managers, and software engineering leaders and executives will also find valuable insights.




The Self-Service Data Roadmap


Book Description

Data-driven insights are a key competitive advantage for any industry today, but deriving insights from raw data can still take days or weeks. Most organizations can’t scale data science teams fast enough to keep up with the growing amounts of data to transform. What’s the answer? Self-service data. With this practical book, data engineers, data scientists, and team managers will learn how to build a self-service data science platform that helps anyone in your organization extract insights from data. Sandeep Uttamchandani provides a scorecard to track and address bottlenecks that slow down time to insight across data discovery, transformation, processing, and production. This book bridges the gap between data scientists bottlenecked by engineering realities and data engineers unclear about ways to make self-service work. Build a self-service portal to support data discovery, quality, lineage, and governance Select the best approach for each self-service capability using open source cloud technologies Tailor self-service for the people, processes, and technology maturity of your data platform Implement capabilities to democratize data and reduce time to insight Scale your self-service portal to support a large number of users within your organization




Future Data and Security Engineering


Book Description

This book constitutes the proceedings of the 7th International Conference on Future Data and Security Engineering, FDSE 2020, which was supposed to be held in Quy Nhon, Vietnam, in November 2020, but the conference was held virtually due to the COVID-19 pandemic. The 24 full papers (of 53 accepted full papers) presented together with 2 invited keynotes were carefully reviewed and selected from 161 submissions. The other 29 accepted full and 8 short papers are included in CCIS 1306. The selected papers are organized into the following topical headings: security issues in big data; big data analytics and distributed systems; advances in big data query processing and optimization; blockchain and applications; industry 4.0 and smart city: data analytics and security; advanced studies in machine learning for security; and emerging data management systems and applications.