Presentation is loading. Please wait.

Presentation is loading. Please wait.

Data Quality. Learning Objectives 1.Identify data quality issues at each step of a data management system 2.List key criteria used to assess data quality.

Similar presentations


Presentation on theme: "Data Quality. Learning Objectives 1.Identify data quality issues at each step of a data management system 2.List key criteria used to assess data quality."— Presentation transcript:

1 Data Quality

2 Learning Objectives 1.Identify data quality issues at each step of a data management system 2.List key criteria used to assess data quality 3.Identify key steps for ensuring data quality at all levels of the data management system

3 Does your Organization have Problems with...... Incomplete or missing data? Getting data collected, collated or analyzed quickly enough? Getting data from partners in a timely manner? Getting honest information?

4 Data Quality: Effect on Program Poor quality data can cause: –Erroneous program management decisions, and the use of additional program resources to take corrective actions –Missed opportunities for identifying program strengths or weaknesses –Reduced stakeholder confidence and support

5 5 Data Management System (DMS) Chain: 6 key processes Source  Collection  Collation  Analysis  Reporting  Use

6 1.Validity – the ability to measure what is intended (accuracy) 2.Reliability – data collected regularly using the same methodology 3.Integrity – data free of “untruth” from willful or unconscious error due to manipulation or through the use of technology (for example data entry) 4.Precision – the ability to minimize random error, or the ability to reproduce measurements consistently 5.Timeliness - ability to have regularly collected, up-to-date data available when it is needed 6.Completeness – all intended is collected (no missing data) Dimensions of Data Quality

7 Information System REAL WORLD In the real world, project activities are implemented in the field. These activities are designed to produce results that are quantifiable INFORMATION SYSTEM An information system represents these activities by collecting the results that were produced and mapping them to a recording system Data Quality: How well the information system represents the real world Data Quality Real World 1. Validity 2. Reliability 3. Integrity 4. Precision 5. Timeliness 6. Completeness Data Quality

8 Case Study 1: Number of Visit Forms Submitted by CHWs, by month DateCHW1CHW2CHW3 Month 1171015 Month 2151214 Month 3161415 Subtotal483644 Total128

9 1. Validity Do the data clearly, directly, and adequately represent the result that was intended to be measured? Have we actually measured what we intended? Do the data adequately represent performance & are thus valid? –Example: Are malaria cases really malaria cases as defined?

10 A Selection of 3 Aspects of Validity 1.Face Validity –Is there a solid, logical relation between the activity or program and what is being measured? –Does the indicator directly measure what was intended to be measured? 2.Measurement Validity –Measurement Errors resulting from poorly designed instruments or poor management of the DMS 3.Transcription Validity –Are transcription / data entry and collation procedures sound? –Is the data entered/ tallied correctly?

11 Data Validity challenges Poorly designed data collection instruments Poorly trained or partisan enumerators Use of questions that elicit incomplete or untruthful information Respondents feel pressured to answer correctly Data are incomplete

12 Data Validity challenges Data altered in transcription Data Manipulation Errors: –Formulae not being applied correctly –Inconsistent formula application –Inaccuracies in formulae Sampling / representation errors –Was the sampling frame wrong? –Unrepresentative Samples: too small for statistical extrapolation and/or contain sample units based on supposition rather than statistical representation

13 Maximizing Validity in your DMS Use well-designed Indicator Reference Sheets to clearly and specifically define what is being measured Use well-designed data collection and collation tools Understand the population you are interested in (ie, sampling frame) Be thoughtful and practical about methodologies (more sophisticated is not always better) Spend a lot of money on training, supervision, and all of the processes and procedures relevant to data collection

14 Case Study 2: Primary School Statistical Profile 2007 DistrictAttendance?Completion?Academic Performance? Berea Butha-Buthe Leribe Mafeteng Maseru Mohale’s Hoek Mokhotlong Qacha’s Nek Quithing Thaba Tseka

15 2. Reliability Consistency and stability of the measurement and data collection process –Can we get the same results or findings if the procedure were repeated over and over? data collection process sampling methodology data collection procedure Three aspects: –Consistency –Internal Quality Control –Transparency

16 Reliability: Consistency Consistent procedures for data collection, maintenance, analysis, processing, and reporting –Definitions and indicators are stable across time –Data collection and sampling processes used are the same across time periods and personnel –Data analysis and reporting approach are the same from year to year, location to location, data source to data source

17 Reliability: Internal Quality Controls Controls are procedures that: –exist to ensure that data are free of significant error & that bias has not been introduced –are in place for periodic review of data collection, maintenance, and processing –provide for periodic sampling and quality assessment of data

18 Reliability: Transparency Transparency is increased when: –Data collection, cleaning, analysis, reporting, & quality assessment procedures are documented in writing –Data problems at each level are reported to the next level –Data quality problems are clearly described in reports

19 Maximize Reliability in your DMS Develop Indicator Protocol Reference Sheets with clear instructions/definitions for all data collection efforts at all levels of the process Document data collection, cleaning, analysis, reporting, and quality assurance procedures Make sure everyone is aware of the procedures and that they follow your procedures from reporting period to reporting period Use the same tool for collecting, collating, and reporting – thereby reducing transcription between DMS stages

20 Maximize Reliability in your DMS Minimize (or eliminate) the use of different tools within and between sites for collecting/collating and reporting the same data Ensure that people report when data is wrong or missing Maintain official copies of all protocols and instruments Consider using multiple (3) independent raters when data is based on human observation

21 Maximize Reliability in your DMS Provide guidance to data collectors and transcribers on what to do if they see a data error and who to report it to Develop an error log to track data errors and document data quality problems Routinely randomly check data for transcription errors. If necessary, double key information Collect data electronically. Include automatic skips/checks in data collection/transcription software Track errors to their original source and correct mistakes Always double check that final numbers are accurate

22 Case Study 3 The data capturer at Organization Z…

23 3. Integrity Data integrity refers to data which is free of “untruth” from willful or unconscious error Compromised during data collection, cleaning, handling and storage due to lack of proper controls by either human or technical means, willfully or unconsciously For example: –loss of data when technology fails Fragile Data Management System –data are manipulated for either personal or political reasons

24 Maximize Integrity Maintain objectivity and independence in data collection, management, and assessment procedures Arrange for more than one data handler/reporter Have a second set of eyes to verify data reported at each stage of the DMS Copy all data handlers in final reports Treat data as “official organizational information” and store data with security Remove incentives that could lead to data manipulation Hire personnel to maintain the technical aspects of the DMS

25 Case Study 4 : # of LLINs distributed by site & month Site 1Site 2Site 3Site 4Total Month 1 20 00025 00023 00024 00072 000 Month 2 2 00023 00024 00022 00071 000 Month 3 23 000 21 00023 00090 000 Total 45 00071 00068 00069 000233 000

26 4. Precision Precision is the ability to minimize random error, or the ability to reproduce measurements consistently Precision is measurement of error “around the edges” of the measure –Margin of error Are the units of measurement refined enough for your application / needs?

27 Maximize Precision Know exactly what kind of information you require Ensure that your data collection staff have the capacity to maintain adequate precision Refine data collection instruments Report any issues around precision (ex. develop an error log)

28 Case Study 5 Organization Green: Organic Pest Control Applications

29 5. Timeliness Two Aspects of Timeliness: –Frequency Data are available on a frequent enough basis to inform program management decisions –Currency The data are reported as soon as possible after collection Main threats to Timeliness concern collection and reporting frequencies

30 Maximize Timeliness Make a realistic schedule that meets program management needs and enforce it! –Include dates for data collection, collation, analysis and reporting- each level of the data management process –Have a calendar for when data is due at each level in the data management process Data Collection/Collation: Clearly identify the date of data collection/collation on all forms and reports Reporting: Clarify dates of data collection when presenting results

31 Case Study 6 Operation Appropriate Treatment: Confirmed Cases and Treatment Maps

32 6. Completeness All of the data that you intend to collect are collected Your DMS represents the entirety of service delivery points, individuals, or other measurement units, and not just a fraction of the full population

33 Remember that Data Quality – Refers to the worth/accuracy of the information collected Focuses on ensuring that the data management process is of a high standard Meets reasonable standards for validity, reliability, timeliness, precision, and integrity

34 Key Factors for Data Quality Success 1. Functioning information systems 2. Clear definition of indicators consistently used at all levels 3. Description of roles and responsibilities at all levels 4. Specific reporting timelines 5. Standard/compatible data-collection and reporting forms/tools with clear instructions 6. Documented data review procedures to be performed at all levels 7. Steps for addressing data quality challenges (missing data, double-counting, lost to follow up, …) 8. Storage policy and filling practices that allow retrieval of documents for auditing purposes (leaving an audit trail)

35 MEASURE Evaluation is a MEASURE program project funded by the U.S. Agency for International Development (USAID) through Cooperative Agreement GHA-A-00-08-00003-00 and is implemented by the Carolina Population Center at the University of North Carolina at Chapel Hill, in partnership with Futures Group International, John Snow, Inc., ICF Macro, Management Sciences for Health, and Tulane University. Visit us online at http://www.cpc.unc.edu/measurehttp://www.cpc.unc.edu/measure


Download ppt "Data Quality. Learning Objectives 1.Identify data quality issues at each step of a data management system 2.List key criteria used to assess data quality."

Similar presentations


Ads by Google