Measuring Data Quality for Ongoing Improvement: A Data Quality Assessment Framework (The Morgan Kaufmann Series on Business Intelligence)

$37.46
by Laura Sebastian-Coleman

Shop Now
The Data Quality Assessment Framework shows you how to measure and monitor data quality, ensuring quality over time. You’ll start with general concepts of measurement and work your way through a detailed framework of more than three dozen measurement types related to five objective dimensions of quality: completeness, timeliness, consistency, validity, and integrity. Ongoing measurement, rather than one time activities will help your organization reach a new level of data quality. This plain-language approach to measuring data can be understood by both business and IT and provides practical guidance on how to apply the DQAF within any organization enabling you to prioritize measurements and effectively report on results. Strategies for using data measurement to govern and improve the quality of data and guidelines for applying the framework within a data asset are included. You’ll come away able to prioritize which measurement types to implement, knowing where to place them in a data flow and how frequently to measure. Common conceptual models for defining and storing of data quality results for purposes of trend analysis are also included as well as generic business requirements for ongoing measuring and monitoring including calculations and comparisons that make the measurements meaningful and help understand trends and detect anomalies. Demonstrates how to leverage a technology independent data quality measurement framework for your specific business priorities and data quality challenges - Enables discussions between business and IT with a non-technical vocabulary for data quality measurement - Describes how to measure data quality on an ongoing basis with generic measurement types that can be applied to any situation "This book provides a very well-structured introduction to the fundamental issue of data quality, making it a very useful tool for managers, practitioners, analysts, software developers, and systems engineers. It also helps explain what data quality management entails and provides practical approaches aimed at actual implementation. I positively recommend reading it…" --ComputingReviews.com, January 2014 "The framework she describes is a set of 48 generic measurement types based on five dimensions of data quality: completeness, timeliness, validity, consistency, and integrity. The material is for people who are charged with improving, monitoring, or ensuring data quality." --Reference and Research Book News, August 2013 "If you are intent on improving the quality of the data at your organization you would do well to read Measuring Data Quality for Ongoing Improvement and adopt the DQAF offered up in this fine book." --Data and Technology Today blog, July 2013 A ready-to-use framework for data quality measurement! The Data Quality Assessment Framework shows you how to measure and monitor data quality, ensuring quality over time. You’ll start with general concepts of measurement and work your way through a detailed framework of more than three dozen measurement types related to five objective dimensions of quality: completeness, timeliness, consistency, validity, and integrity. Ongoing measurement, rather than one time activities will help your organization reach a new level of data quality. This plain-language approach to measuring data can be understood by both business and IT and provides practical guidance on how to apply the DQAF within any organization enabling you to prioritize measurements and effectively report on results. Strategies for using data measurement to govern and improve the quality of data and guidelines for applying the framework within a data asset are included. You’ll come away able to prioritize which measurement types to implement, knowing where to place them in a data flow and how frequently to measure. Common conceptual models for defining and storing of data quality results for purposes of trend analysis are also included as well as generic business requirements for ongoing measuring and monitoring including calculations and comparisons that make the measurements meaningful and help understand trends and detect anomalies. Laura Sebastian-Coleman, Data Quality Director at Prudential, has been a data quality practitioner since 2003. She has implemented data quality metrics and reporting, launched and facilitated working stewardship groups, contributed to data consumer training programs, and led efforts to establish data standards and manage metadata. In 2009, she led a group of analysts in developing the Data Quality Assessment Framework (DQAF), which is the basis for her 2013 book, Measuring Data Quality for Ongoing Improvement. An active professional, Laura has delivered papers, tutorials, and keynotes at data-focused conferences, such as MIT’s Information Quality Program, Data Governance and Information Quality (DGIQ), Enterprise Data World (EDW), Data Modeling Zone, and Data Management Association (DAMA)-sponsored events. From 2009 to 2010, she served as

Customer Reviews

No ratings. Be the first to rate

 customer ratings


How are ratings calculated?
To calculate the overall star rating and percentage breakdown by star, we don’t use a simple average. Instead, our system considers things like how recent a review is and if the reviewer bought the item on Amazon. It also analyzes reviews to verify trustworthiness.

Review This Product

Share your thoughts with other customers