Data lifecycle Data Management & Survey Conception

Slides:



Advertisements
Similar presentations
Presented by: Denise Sjahkit SURINAME. Introduction Overview of the main policy issues Scope Current compilation practices Data-sources Requirements for.
Advertisements

Tips and Resources IASC Cluster/Sector Leadership Training
DS-01 Disaster Risk Reduction and Early Warning Definition
1 Measuring ICT4D: ITUs Focus on Household and Individual Market, Economics & Finance Unit Telecommunication Development Bureau.
Evaluation Capacity Building Identifying and Addressing the Fields Needs.
Supporting National e-Health Roadmaps WHO-ITU-WB joint effort WSIS C7 e-Health Facilitation Meeting 13 th May 2010 Hani Eskandar ICT Applications, ITU.
Global Congress Global Leadership Vision for Project Management.
Near East Plant Protection Network for Regional Cooperation & Knowledge Sharing Food and Agriculture Organization of the United Nations An Overview on.
C6 Databases.
PARIS 21 Meeting Ghana July Challenges in health information Health Metrics Network HMN Framework and profiling tool.
Comprehensive M&E Systems
Statistical Metadata Strategy Elham M. Saleh - Acting Director of Economic Statistics - Director of Technical Resources Central Informatics Organisation.
Improving the Energy Efficiency of the Heat and Hot Water Supply Presenter: Bayramgul Garabaeva, Programme Officer Decentralization and Community Development.
Distributed Data Analysis & Dissemination System (D-DADS) Prepared by Stefan Falke Rudolf Husar Bret Schichtel June 2000.
System of Environmental-Economic Accounting SEEA Implementation Guide and Diagnostic Tool Alessandra Alfieri UNSD.
Magdi Latif Regional Knowledge and Information Management Officer FAO Partnership, Advocacy and Capacity Development Division FAORNE Jordan Plant Genetic.
NGAC Interagency Data Sharing and Collaboration Spotlight Session: Best Practices and Lessons Learned Robert F. Austin, PhD, GISP Washington, DC March.
Open Development Landscape in Uganda Uganda Open Development Stakeholders Workshop, Hotel Africana 11 th - 12 th September 2012.
FSC Information Management Tool - introduction. Training Overview What is the FSC? What is Information Management? Why a FSC IM Tool? FSC Reporting Reporting.
1 Capacity Building: Strategy and Action Plan GEF-UNDP Strategic Partnership Capacity Development Initiative.
Global Action Plan and its implementation in other regions Meeting for Discussion of the draft Plan for the Implementation of the Global Strategy to Improve.
ITEC224 Database Programming
Global Strategy to Improve Agricultural Statistics Food and Agriculture June 22, 2009 Organization.
Regional Integration Cluster Olivier Hartmann - SSATP.
Achievements to date on the IDP Profiling Process UNFPA UNHCR Brussels IDP Profiling Workshop 22 – 23 March 2007.
Data Warehouse Overview September 28, 2012 presented by Terry Bilskie.
Assessments. Assessment in the Project Cycle DESIGN IMPLEMENTATION MONITORING EVALUATION ASSESSMENT.
IODE Ocean Data Portal - technological framework of new IODE system Dr. Sergey Belov, et al. Partnership Centre for the IODE Ocean Data Portal MINCyT,
ROLE OF INFORMATION IN MANAGING EDUCATION Ensuring appropriate and relevant information is available when needed.
Assessing The Development Needs of the Statistical System NSDS Workshop, Trinidad and Tobago, July 27-29, 2009 Presented by Barbados.
A UNHCR Perspective on IDP Profiling Interagency Workshop Brussels 22 – 23 March 2007.
Challenges in Data Collection and Dissemination of National Development Indicators, including MDGs The Experience of Cambodia by Lay Chhan National Institute.
C6 Databases. 2 Traditional file environment Data Redundancy and Inconsistency: –Data redundancy: The presence of duplicate data in multiple data files.
Africa Health Workforce Platform & Observatory Presentation to the 1st conference of the Asia-Pacific Action Alliance on HRH (AAAH): October 2006.
10/10/2012ISC239 Isabelle Bichindaritz1 Physical Database Design.
THE UNITED REPUBLIC OF TANZANIA Millennium Development Goals (MDGs) Monitoring Workshop Kampala Uganda, 5 th - 8 th May 2008 COORDINATION OF NATIONAL STATISTICAL.
Statistics New Zealand’s End-to-End Metadata Life-Cycle ”Creating a New Business Model for a National Statistical Office if the 21 st Century” Gary Dunnet.
United Nations Economic Commission for Europe Statistical Division Mapping Data Production Processes to the GSBPM Steven Vale UNECE
Opportunity & challenge Nampung Chirdchuepong National Statistical Office Thailand Nampung Chirdchuepong National Statistical Office Thailand
Looking Forward the Post-2015 Sustainable Development Goals (SDGs) UNSD-DFID Project on National Development Indicators: Country Director's Meeting, New.
Pilot Projects on Strengthening Inventory Development and Risk Management-Decision Making for Mercury: A Contribution to the Global Mercury Partnership.
Draft GEO Framework, Chapter 6 “Architecture” Architecture Subgroup / Group on Earth Observations Presented by Ivan DeLoatch (US) Subgroup Co-Chair Earth.
Indicators to Measure Progress and Performance IWRM Training Course for the Mekong July 20-31, 2009.
African Centre for Statistics United Nations Economic Commission for Africa Data compilation and management proposal for Africa Statistical Commission.
Consultant Advance Research Team. Outline UNDERSTANDING M&E DATA NEEDS PEOPLE, PARTNERSHIP AND PLANNING 1.Organizational structures with HIV M&E functions.
Observatories for the Health Workforce in Africa VIII REGIONAL MEETING OF THE OBSERVATORIES OF HUMAN RESOURCES FOR HEALTH IN THE AMERICAS LIMA, PERU
United Nations Oslo City Group on Energy Statistics OG7, Helsinki, Finland October 2012 ESCM Chapter 8: Data Quality and Meta Data 1.
Distributed Data Analysis & Dissemination System (D-DADS ) Special Interest Group on Data Integration June 2000.
Copyright 2010, The World Bank Group. All Rights Reserved. Managing processes Core business of the NSO Part 1 Strengthening Statistics Produced in Collaboration.
EO Dataset Preservation Workflow Data Stewardship Interest Group WGISS-37 Meeting Cocoa Beach (Florida-US) - April 14-18, 2014.
Modernising Statistical Production: Modernising Statistical Production: Main recommendations from global assessments 7 th SPECA PWG on Statistics
Registration and Profiling Haiti Inter-Cluster Meeting 10 February 2010.
5.8 Finalise data files 5.6 Calculate weights Price index for legal services Quality Management / Metadata Management Specify Needs Design Build CollectProcessAnalyse.
Comprehensive M&E Systems: Identifying Resources to Support M&E Plans for National TB Programs Lisa V. Adams, MD E&E Regional Workshop Kiev, Ukraine May.
Development Account: 6th Tranche Strengthening the capacity of National Statistical Offices (NSOs) in the Caribbean Small Island Developing States to fulfill.
IODE Ocean Data Portal - technological framework of new IODE system Dr. Sergey Belov, et al. Partnership Centre for the IODE Ocean Data Portal.
Criteria for Assessing MHPSS Proposals Submitted through the CAP, CERF and HRF Funding Mechanisms to the Protection Cluster.
System Overview Training on the use of the new countrystat
Session 2: Institutional arrangements for energy statistics
12. Role of national and international technical and funding partners in the implementation of aDSM Multi-partner training package on active TB drug safety.
System Overview Training on the use of the new countrystat
System Overview Training on the use of the new countrystat
Using the Checklist for SDMX Data Providers
Data Warehouse Overview September 28, 2012 presented by Terry Bilskie
C.U.SHAH COLLEGE OF ENG. & TECH.
Albania 2021 Population and Housing Census - Plans
Data Warehousing Concepts
Mapping Data Production Processes to the GSBPM
Comprehensive M&E Systems
Palestinian Central Bureau of Statistics
Presentation transcript:

Data lifecycle Data Management & Survey Conception Data System Architecture Data collection management Data Analysis & Dissemination 1 1

Type of info per usage Introduction 2

From Data.. to Information Introduction Operation Data Manager should be involved in all the steps of a “Data Lifecycle”. Any break of this cycle ends with the failure of the system : A data collection form that is ill- designed either because it does not satisfy operational information requirements or is flawed from a technical standpoint A well designed survey with a poorly designed and therefore poorly maintained database A structurally well designed database with no data, as data collection cycles have not been integrated/respected A well populated database without implemented reports and queries and therefore no output Survey Conception Data System Architecture Collection management Data Analysis & Dissemination 3

Layers of data collection Before the Form… Survey Conception Avoid reinventing the wheel – check what has been designed and piloted before Consultation with all stakeholders – avoid duplication of efforts and assessment fatigue of beneficiaries Layers of data collection Collect Simple base reference data first Embark on detailed info based on samples defined from the base reference Data collection frequency should vary according to how frequently the phenomena being tracked or measured changes 4

Good practices for Data Collection Forms Survey Conception Good practices for Data Collection Forms Questionnaires used in survey research should be clear and well presented. Think about the form of the questions, Keep the survey as short as possible. Make definitions of data elements consistent with standard definitions and analytic conventions Plan clearly how answers will be analyzed. Test the survey for “understandability” and respondent effort through focus groups Good practices for Data Collection Forms Questionnaires used in survey research should be clear and well presented. The use of capital (upper case) letters only should be avoided, as this format is hard to read. Questions should be numbered and clearly grouped by subject, each respondent being asked the same questions in the same order. Clear instructions should be given and headings included to make the questionnaire easier to follow. Provide definitions and clarifying information for individual items and terms. Be clear about the date or time period the survey should reflect. Be sure that return address, fax number, and contact information are included on the form itself as well as in accompanying letter. Consider including an addressed, postage-paid envelope to return the survey. Think about the form of the questions, Avoid ‘double-barreled’ questions (two or more questions in one, ie. do not combine two separate ideas inappropriately and request a single response). Avoid questions containing double negatives, and leading or ambiguous questions. Use standard language. Avoid jargon and abbreviations If closed questions are used the interviewer should also have a range of pre-coded responses available. Make categories as concrete as possible (Example: Replace vague quantifying words (e.g., frequently, often, sometimes) with specific numerical ranges whenever possible. If response categories are supposed to be mutually exclusive and exhaustive, make sure all possible categories are included on the survey. If in doubt about whether categories are exhaustive, include a response option of "other, please specify." Indicate subpopulations to include and/or exclude in the reporting. If data may be unknown or missing, be sure to include an “unknown/missing” option. Provide a “not applicable” response for questions that may not be applicable to all respondents. Design the item sequence of the survey such that it increases the respondent's ability to complete the survey. Keep topic-related questions together and provide transitions between topics. Fro example, each type of data is listed in a separate section with a clear heading that identifies the topic (e.g., general information, demographic, population movement). Definitions and cohorts within one survey and across related surveys should be consistent. If data for a cohort are reported in more than one section of the survey, indicate that grand totals for each section should match. Example: total by age should equal total by ethnicity. Keep the survey as short as possible. Ensure that the response burden does not exceed a reasonable length and is justified by the use of the data. Examine each item in the data collection instrument to make sure that the information is needed and will be used. Avoid requesting information that is of marginal use. Avoid requesting data that can be obtained from another available survey or database (Example: Do not ask for disaggregating data that will not be used). Make definitions of data elements consistent with standard definitions and analytic conventions (i.e., calculations or methodologies) when appropriate and feasible. If appropriate, use definitions that conform to definitions developed nationally to ensure that the data reported will be comparable to data reported by other agencies and organizations at the institutional, state, and federal levels. If standard definitions and/or analytic conventions are used, indicate the sources of the definitions or conventions used. A number of resources are identified in the appendices. Determine whether another organization is already collecting data related to the items you plan to collect; if so, obtain a copy of that survey and consider using the same definitions and analytic conventions as a starting point for your survey. Another option is to ask respondents to report data directly from the other survey or to request a copy of the completed survey. Any deviations from accepted usage should be explained. If more than one analytic convention is commonly used for a given measure, indicate the preferred convention. Ask respondents to indicate if they used a different analytic convention and why. Questions may be open (where the respondent composes the reply) or closed (where pre-coded response options are available, e.g. multiple-choice questions). Whether using open or closed questions, data managers should plan clearly how answers will be analyzed. Closed questions with pre-coded response options are most suitable for topics where the possible responses are known. Closed questions are quick to administer and can be easily coded and analyzed. Open questions should be used where possible replies are unknown or too numerous to pre-code. Open questions are more demanding for respondents but if well answered can provide useful insight into a topic. Open questions, however, can be time consuming to administer and difficult to analyze. Open questions are used more frequently in unstructured interviews, whereas closed questions typically appear in structured interview schedules. A structured interview is like a questionnaire that is administered face to face with the respondent. Whenever possible, test the survey for “understandability” and respondent effort through focus groups, cognitive laboratory sessions, or pilot testing. Whenever possible, conduct a “trial-run” in which results are made available only to survey respondents before releasing them publicly. The purpose of these activities is to ensure that: Each item is understandable to the respondent The technical terms used are appropriate to the respondent The questions are clear and unambiguous to the respondent The items elicit a single response The survey is not too much of a burden for the respondent Seek periodic review of the survey instrument from experienced, knowledgeable individuals.

Data model Data System Architecture Data models are the key for interoperability (i.e easy data exchange with partners) Implementing partners should not have to draft and decide on a core data model; it should be the same everywhere and just adapted locally where necessary; support (guidelines) need to be there Importance of a common referential Site / community Assessment Beneficiary registration Demographics Multi sectoral assessment: Health Education Water Bio Data Vulnerability Needs Site Infrastructure Inventory Base indicators Delivered Assistance Organization Who’s doing what where? Project activities description Performance Indicators Activity monitoring 6

Building an Interface for data collection: System architecture Data System Architecture Building an Interface for data collection: Mobile Offline desktop Web/Server based OCR* ready form (can be scanned) Integration of external data source (ETL**) Offering analysis capacity (OLAP*** and Stats) * Mechanical or electronic translation of scanned images of handwritten, typewritten or printed text into machine-encoded text ** Extract, transform, and load (ETL) is a process in database usage that involves Extracting data from outside sources, Transforming it to fit operational needs (which can include quality levels), Loading it into the end target (database or data warehouse) *** An OLAP (Online analytical processing) cube is a data structure that allows fast analysis of data. 7

Reports are part of the data system Data System Architecture Queries and tools to extract data from the databases need to be designed along with the database Must give abilities for reporting officers to Set up queries and reports without high level IT knowledge To be clear on the standard indicators these queries should be based on 8

Data collection strategies Data collection management Direct coordination with partners ex : Somali protection cluster Establishment of a « data collection project » ex : UNOPS Goma Specific Contract with a dedicated partner Ex: CartONG in Uganda 9

Avoid conflict of interest Implementation matrix Data collection management Avoid conflict of interest 10

Targets mostly local partners and decision makers PDF reports and maps Data Dissemination Targets mostly local partners and decision makers Can be disseminated through mailing list (cf Somali protection) Google group (cf Goma Update) Website (cf ReliefWeb) 11

GeoPortal and Open Data API Data Dissemination GeoPortal: is a tool to ensure institutional memory and “Master Data” management Can be a tool for desk officers to visualize a situation and use map extracts in their reporting Data API: Can be used for global dissemination: cf Worldbank Data API or Google public data Offers material for data journalism (e.g. computer assisted reporting on data through journalists) 12

Data, Law & License Data Dissemination For all data sets that do not fall under the “Guidelines for the Regulation of Computerized Personal Data Files” (for instance protection data) …. …. The “Open database license” (ODBL) can give a legal frame to all our data collection activities http://www.unhcr.org/refworld/pdfid/3ddcafaac.pdf http://www.opendatacommons.org/licenses/odbl/1.0 13

Providing support for the 4 phases of the process Conclusion 4 specific types of expertise that are difficult to combine in one profile: Statistician/Analyst: Creating a questionnaire and compiling analyzing the resulting statistics IS Architect: Building the information system Manager: Managing the stakeholder consultation process during the design phase, the collection in the field and dissemination of results Data journalist: Developing sound and sexy reports Need to find where are the gap among the “Operation Data Management” officers network Need to define the training & support need for each of those specific domains 14