DEPT, FERRARI AND MENDELOVITS: HOW TO ANALYZE AND EXPLOIT FIELD TEST RESULTS WITH A VIEW TO MAXIMIZING CROSS-LANGUAGE COMPARABILITY OF MAIN SURVEY DATA.

Slides:



Advertisements
Similar presentations
ICP 7-th Regional Coordinators Meeting World Bank, Washington D.C.
Advertisements

International Reading Association
Principal Investigators: Martha Thurlow & Deborah Dillon Introduction Assumptions & Research Questions Acknowledgments 1. What characteristics of current.
Internet Survey Method in the 2010 Census and Challenges to the 2015 Census in Japan Population Census Division, Statistics Bureau of Japan Hideki Koizumi.
PIAAC: Delivering the Survey in Ireland
Translation and Translation Verification National Research Coordinators Meeting Amsterdam, October 2006.
Translation and Verification of ICCS instruments Summary of the Field Trial & Plans for the Main Study National Research Coordinators Meeting Windsor,
PISA FOR DEVELOPMENT Technical Workshops Components and input for ToR of the International Contractor(s) 9 th April 2014 OECD Secretariat 1.
Chapter 7: Key Process Areas for Level 2: Repeatable - Arvind Kabir Yateesh.
Overview of field trial analysis procedures National Research Coordinators Meeting Windsor, June 2008.
Item Analysis Prof. Trevor Gibbs. Item Analysis After you have set your assessment: How can you be sure that the test items are appropriate?—Not too easy.
Short Course on Introduction to Meteorological Instrumentation and Observations Techniques QA and QC Procedures Short Course on Introduction to Meteorological.
Chapter 13 Survey Designs
Creating Research proposal. What is a Marketing or Business Research Proposal? “A plan that offers ideas for conducting research”. “A marketing research.
Proposal Writing.
Sales and Marketing Productivity Team 1 Added Value Analysis TOOL USED IN SALES AND MARKETING PRODUCTIVITY PROJECTS.
Conducting a Job Analysis to Establish the Examination Content Domain Patricia M. Muenzen Associate Director of Research Programs Professional Examination.
Section 29.2 The Marketing Survey
Power Point Slides by Ronald J. Shope in collaboration with John W. Creswell Chapter 13 Survey Designs.
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 24 Slide 1 Critical Systems Validation 1.
1st NRC Meeting, October 2006, Amsterdam 1 ICCS 2009 Field Operations.
Assessing higher education learning outcomes globally Professor Hamish Coates
Marketing Research  Def. - Formal communication link with the environment to provide accurate and useful information for better decision making.  Systematic.
XLIFF in a large-scale international OECD study Case study Britta Upsing, Steve Dept, Andrea Ferrari, Heiko Rölke.
Chapter 2 Analyzing Orgs and Jobs Organizational need analysis chapter 2 Analyzing Orgs and Jobs1.
Data Management Seminar, 8-11th July 2008, Hamburg ICCS 2009 Main Survey Field Operations.
Streamlining the Review Cycle Michael Oettli, nlg GmbH Santa Clara, October 10 th.
Action Research March 12, 2012 Data Collection. Qualities of Data Collection  Generalizability – not necessary; goal is to improve school or classroom.
SPECA Regional Wrokshop on Disability Statistics, Dec 13-15, 2006 Issues Related to Disability Measurement: Cognitive testing and mode Jennifer Madans.
ESS Slide 1 Quality assessment of MEHM in SILC Eurostat Unit F5 “Health and Food Safety Statistics” 3 rd meeting of the Task Force on Health Expectancies.
1 U.S. PIAAC National Supplement: Prison Study Overview Association of State Correctional Administrators Research and Best Practices Committee Gaylord.
1 Chapter 2 The Process. 2 Process  What is it?  Who does it?  Why is it important?  What are the steps?  What is the work product?  How to ensure.
Early Childhood Development (ECD) Scales: Overview & Lessons Learned Dr. Ghazala Rafique Aga Khan University Pakistan Regional Consultative Meetings on.
Writing research proposal/synopsis
Evaluating a Research Report
Measuring of student subject competencies by SAM: regional experience Elena Kardanova National Research University Higher School of Economics.
3rd NRC Meeting, 9-12 June 2008, Windsor ICCS 2009 Main Survey Field Operations.
Assessing assessment: the role of student effort in comparative studies Ray Adams Jayne Butler.
National adaptations to main survey instruments and layout verification National Research Coordinators Meeting Windsor, June 2008.
Using Data in the Goal-Setting Process Webinar September 30, 2015.
By Cao Hao Thi - Fredric W. Swierczek
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
Perceptive Agile Measurement: New Instruments for Quantitative Studies in the Pursuit of the Social-Psychological Effect of Agile Practices Department.
Data Management Seminar, 9-12th July 2007, Hamburg 11 ICCS 2009 – Field Trial Survey Operations Overview.
Israel Accession Seminar PIAAC: Programme for International assessment of Adult Competencies Skills strategy in OECD Programme for the International Assessment.
Data Management Seminar, 8-11th July 2008, Hamburg 1 ICCS 2009 – On-line Data Collection in the Main Survey.
Automated Assessment Management System. The Assessment Cycle Trainee | Learner Dashboard Trainer Dashboard Employer Dashboard Verifier Dashboard Assessor.
Question paper 1997.
United Nations Oslo City Group on Energy Statistics OG7, Helsinki, Finland October 2012 ESCM Chapter 8: Data Quality and Meta Data 1.
1 Cognitive Aspects Associated with Sample Selection Conducted by Respondents in Establishment Surveys La Toya Barnett Thomas Rebecca L. Morrison Grace.
Margaret Blake RESEARCH METHODS FESTIVAL: Evaluating questionnaires with cognitive testing.
Qualifications Update: Higher Media Qualifications Update: Higher Media.
Language Testing How to make multiple choice test.
Revised OSDV Tools and Processes LFA PSM expert workshop 28-30January 2014.
Assistant Instructor Nian K. Ghafoor Feb Definition of Proposal Proposal is a plan for master’s thesis or doctoral dissertation which provides the.
11 PIRLS The Trinidad and Tobago Experience Regional Policy Dialogue on Education 2-3 December 2008 Harrilal Seecharan Ministry of Education Trinidad.
Assessment and the Institutional Environment Context Institutiona l Mission vision and values Intended learning and Educational Experiences Impact Educational.
By Dr Hidayathulla Shaikh.  At the end of the lecture student should be able to -  mention steps in methodology of a study  Discuss steps in methodology.
The Delivery Matters: Examining Reactivity in Question Answering Riley E. Foreman, Marshall T. Beauchamp, and Erin M. Buchanan Missouri State University.
ESRA- 6 th Annual Conference Reykjavik, July 17, 2015 Spoken language versus written language: A challenge for the linguistic validation.
Survey Research and Methodology UNL-GALLUP Research Center Evaluating Data quality in Time Diary Surveys Using Paradata Ana Lucía Córdova Cazar Robert.
WHOLE CHILD- WHOLE SCHOOL One hundred years from now it will not matter what your bank account was, the sort of house you lived in, or the kind of car.
ARDHIAN SUSENO CHOIRUL RISA PRADANA P.
Method.
Classroom Analytics.
Week 8 Slides.
Lies, Damned Lies & Statistical Analysis for Language Testing
Introduction and Objectives Addis Ababa, Ethiopia, December 2012
PISA for Development Technical Workshops Contextual questionnaires
Quality Management Anita Téringer– ITS Hungary
Presentation transcript:

DEPT, FERRARI AND MENDELOVITS: HOW TO ANALYZE AND EXPLOIT FIELD TEST RESULTS WITH A VIEW TO MAXIMIZING CROSS-LANGUAGE COMPARABILITY OF MAIN SURVEY DATA COLLECTION INSTRUMENTS Washington D.C., March 23, 2012

State of the Art  Proliferation of multilingual comparative studies  Survey design includes a pilot, or Field Test (FT) carried out on a smaller scale  items adapted in multiple languages before FT  key moment for linguistic quality control (LQC) : right before FT – translation verification  Comprehensive documentation of adaptation, adjudication and validation processes

Between Field Test and Main Survey  In survey designs that include FT and MS, analysis of FT results = a wealth of information  Can be used to inform item selection  But also to perform a more focused linguistic and formal verification before MS  Open communication channels between item writers, national experts, verification coordinators

The PISA paradigm  Inception in 2000, currently 5 th survey cycle  Double translation, double source design  32 national versions (2000) -> 85 n. v. (2012)  from pencil and paper to computer-delivered assessments and background questionnaires  compiling data on adaptation history of each item in each language

Analysis of FT Results  At item level: item stats (itanals) Item discrimination Item fit Ability ordering Point biserial correlation (MCQ)  Differential item analysis gender country language

Multiple choice item: not dodgy Options A,B,C,D Key Answer Should be positive for key answer Should be negative for distractor Mean ability and standard deviation for the group of students who selected responses A, B, C or D Higher than 0.2 Item fit

Multiple choice item: Dodgy Less than 0.2 Low discrimination between high and low achiever Value significantly higher than 1 (item discrimination between high and low achievers is less than expected)

Action  Dodgy item reports sent to countries to cApStAn  reflect on the data, examine national version; explain why such results may have occurred.  As a result, FT to MS corrections proposed by: Item writers / test developers countries / verifiers

Dodgy item

MS version management  Base national MS version prepared for countries (using final FT version) segment status indicates type of action locked segments if no FT > MS changes  Country review followed by focused verification  Difference reports (before/after) generated automatically  Reports examined by referee  Final check on key corrections

CHALLENGES  Convincing reviewers / verifiers that if it isn’t broken, don’t fix it  Document each change with its justification  Check whether changes have not introduced new errors or inconsistencies  Make more systematic use of dodgy item reports, including for background questionnaires  Embed these processes in the platforms and IT adaptation management systems

ANY QUESTIONS? THANK YOU