Www.rti.org RTI International is a registered trademark and a trade name of Research Triangle Institute. Using Paradata for Monitoring Interviewer Performance.

Slides:



Advertisements
Similar presentations
Correcting for Common Causes of Nonresponse and Measurement Error Andy Peytchev International Total Survey Error Workshop Stowe, June 15, 2010.
Advertisements

Progress Monitoring project DATA Assessment Module.
1 Sampling Telephone Numbers and Adults, and Interview Length, and Weighting in the California Health Survey Cell Phone Pilot Study J. Michael Brick, Westat.
Microsimulation of Survey Collection Yves Bélanger Kristen Couture 26 January 2010.
Enhancing Data Quality of Distributive Trade Statistics Workshop for African countries on the Implementation of International Recommendations for Distributive.
Multiple Indicator Cluster Surveys Survey Design Workshop
RTI International is a trade name of Research Triangle Institute th Street, NW ■ Suite 750 ■ Washington, DC, USA Phone
Elements of a clinical trial research protocol
Challenges in Conducting Multi-Center Clinical Studies: Results from the Rapid Empiric Treatment with Oseltamivir Study (RETOS) Kendra Thompson, Kelly.
Chapter 7 Sampling Distributions
Non-Experimental designs: Developmental designs & Small-N designs
Documentation and survey quality. Introduction.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
FINAL REPORT: OUTLINE & OVERVIEW OF SURVEY ERRORS
08/08/2015 Statistics Canada Statistique Canada Paradata Collection Research for Social Surveys at Statistics Canada François Laflamme International Total.
Robert delMas (Univ. of Minnesota, USA) Ann Ooms (Kingston College, UK) Joan Garfield (Univ. of Minnesota, USA) Beth Chance (Cal Poly State Univ., USA)
How to Assess Student Learning in Arts Partnerships Part II: Survey Research Revised April 2, 2012 Mary Campbell-Zopf, Ohio Arts Council
© 2015 albert-learning.com Marketing Research Process MARKETING RESEARCH PROCESS.
Chapter 4 Principles of Quantitative Research. Answering Questions  Quantitative Research attempts to answer questions by ascribing importance (significance)
Power Point Slides by Ronald J. Shope in collaboration with John W. Creswell Chapter 13 Survey Designs.
Power Point Slides by Ronald J. Shope in collaboration with John W. Creswell Chapter 13 Survey Designs.
Qualitative Studies: Case Studies. Introduction l In this presentation we will examine the use of case studies in testing research hypotheses: l Validity;
The Role of Local Specificity in the Interpretation of Small Area Estimation Benmei Liu Scott Gilkeson Gordon Willis Rocky Feuer 2012 FCSM Statistical.
Presented at The 129th Annual Meeting of the American Public Health Association Atlanta, GA, October 21–25, 2001 Presented by Amanda Honeycutt Abigail.
Modeling errors in physical activity data Sarah Nusser Department of Statistics and Center for Survey Statistics and Methodology Iowa State University.
Single-Case Research: Standards for Design and Analysis Thomas R. Kratochwill University of Wisconsin-Madison.
Responsive Design for Household Surveys: Illustration of Management Interventions Based on Survey Paradata Robert M. Groves, Emilia Peytcheva, Nicole Kirgis,
FAEIS Project User Opinion Survey 2005 Thursday, June 23, 2005 Washington, D.C. H. Dean SutphinYasamin Miller ProfessorDirector, SRI Agriculture & Extension.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
A Latent Class Call-back Model for Survey Nonresponse Paul P. Biemer RTI International and UNC-CH Michael W. Link Centers for Disease Control and Prevention.
The 2006 National Health Interview Survey (NHIS) Paradata File: Overview And Applications Beth L. Taylor 2008 NCHS Data User’s Conference August 13 th,
Deliverable 2.6: Selective Editing Hannah Finselbach 1 and Orietta Luzi 2 1 ONS, UK 2 ISTAT, Italy.
Major Types of Quantitative Studies Descriptive research –Correlational research –Evaluative –Meta Analysis Causal-comparative research Experimental Research.
1 Renee M. Gindi NCHS Federal Conference on Statistical Methodology Statistical Policy Seminar December 4, 2012 Responsive Design on the National Health.
A Strategy for Prioritising Non-response Follow-up to Reduce Costs Without Reducing Output Quality Gareth James Methodology Directorate UK Office for National.
Stop the Madness: Use Quality Targets Laurie Reedman.
CDIS 5400 Dr Brenda Louw 2010 Validity Issues in Research Design.
Single-Subject Experimental Research
The Challenge of Non- Response in Surveys. The Overall Response Rate The number of complete interviews divided by the number of eligible units in the.
Training Individuals to Implement a Brief Experimental Analysis of Oral Reading Fluency Amber Zank, M.S.E & Michael Axelrod, Ph.D. Human Development Center.
Evaluating Ongoing Programs: A Chronological Perspective to Include Performance Measurement Summarized from Berk & Rossi’s Thinking About Program Evaluation,
EDCI 696 Dr. D. Brown Presented by: Kim Bassa. Targeted Topics Analysis of dependent variables and different types of data Selecting the appropriate statistic.
Testing for Coverage Bias when Combining Directory-Listed And Cellphone Samples T. M. Guterbock, A. Diop, J. M. Ellis, J. L. P. Holmes and K. T. Le, Center.
Applying impact evaluation tools A hypothetical fertilizer project.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
PLCS & THE CONNECTION TO RESPONSE TO INTERVENTION Essentials for Administrators Sept. 27, 2012.
1 Collecting Data in Multiple Languages: Development of a Methodology SLAITS National Survey of Children with Special Health Care Needs Seth T. Brady,
1 Quality Control for Field Operations. 2 Overview Goal To ensure the quality of survey field work Purpose To detect and deter interviewer errors and.
11 How Much of Interviewer Variance is Really Nonresponse Error Variance? Brady T. West Michigan Program in Survey Methodology University of Michigan-Ann.
TQM in Interview Data Collection Process Matti Simpanen.
Paradata and Survey Management in the National Survey of Family Growth (NSFG), William D. Mosher, NCHS; James Wagner, Mick Couper, & Nicole Kirgis,
1 Responsive Design and Survey Management in the National Survey of Family Growth (NSFG) William D. Mosher, NCHS FCSM Statistical Policy Seminar Washington,
Basic Business Statistics, 8e © 2002 Prentice-Hall, Inc. Chap 1-1 Inferential Statistics for Forecasting Dr. Ghada Abo-zaid Inferential Statistics for.
Response rates and disposition coding PHC 6716 June 8, 2011 Chris McCarty.
4.1 Statistics Notes Should We Experiment or Should We Merely Observe?
1 Collecting and Interpreting Quantitative Data Deborah K. van Alphen and Robert W. Lingard California State University, Northridge.
Identify bottlenecks Facilitators’ Workshop on District Health Performance Improvement Lilongwe, 25 th – 27 th March 2015.
THE MARKETING RESEARCH PROCESS CHAPTER 29 Mrs. Simone Seaton Marketing Management.
Crystal Reinhart, PhD & Beth Welbes, MSPH Center for Prevention Research and Development, University of Illinois at Urbana-Champaign Social Norms Theory.
Project VIABLE - Direct Behavior Rating: Evaluating Behaviors with Positive and Negative Definitions Rose Jaffery 1, Albee T. Ongusco 3, Amy M. Briesch.
TOPIC - Page 1 Representativity Indicators for Survey Quality R-indicators and fieldwork monitoring Koen Beullens & Geert Loosveldt K.U.Leuven.
UTILIZING TELEPHONE INTERVIEWERS AS COUNSELORS: LESSONS LEARNED FROM A SMOKING REDUCTION STUDY Bridget Gaglio, MPH 1, Tammy Smith, BS 2, Erica Ferro, MA.
Research Methods & Design Outline
Best Practices for Implementing a Paradata Warehouse Presentation at the Washington Statistical Society Mini-Conference on Paradata Washington, D.C. Jason.
Modifying Interviewer Strategies to Reduce Cost of Data Collection
پرسشنامه کارگاه.
Assessing Quality of Paradata to Better Understand the Data Collection Process for CAPI Social Surveys François Laflamme Milana Karaganis European Conference.
The Language of Sampling
Basic Elements of a Graph
Presentation transcript:

RTI International is a registered trademark and a trade name of Research Triangle Institute. Using Paradata for Monitoring Interviewer Performance in Telephone Surveys Emilia Peytcheva, RTI International Andy Peytchev, University of Michigan

Acknowledgements  Howard Speizer  Kelly Castleberry  Tamara Terry  T.J. Nesius  Marion Schultz  Marcus Berzofsky

Outline  Why model paradata  Present objective  Approach in centralized call center setting  History – 2010 implementation  Statistical model  Graphing  Tabulation – randomized experiment – current and future work to enrich the set of metrics

Paradata  Raw paradata have limited utility – Raw: Number of interviews from the full sample – Modelled: AAPOR RR#3 – Raw: Daily interviewing hours – Modelled: Daily hours per interview, by interviewer

Objective  Interviewer administered surveys rely extensively on the ability to: – Identify lowest performing interviewers and take corrective actions – Do so soon after the interviewer starts working on the study

Interviewer Performance Monitoring  Supervisors track hours per complete, response rates, refusal rates, and other similar metrics by interviewer  Even in centralized call center settings, sample cases are not randomly assigned to interviewers  Current practice: use “raw” paradata  Supervisor have to take into account that there may be alternative explanations for an interviewer’s poor performance measures, e.g.: – She worked difficult shifts – She was experienced and therefore assigned to work prior refusals – She worked Spanish language cases

Model-Based Interviewer Performance  Key objective: rate of obtaining interviews  Usual metric: hours per interview  Model-based alternative: ratio of the rate of obtaining interviews to the expected rate of obtaining interviews

Expected Rate of Obtaining Interviews  Estimate the likelihood of obtaining an interview on each call attempt and aggregate to the interviewer level  Control for major departures from random assignment of cases to interviewers – Time of day and day of week (i.e., shift) – Prior refusal (i.e., refusal queue) – Appointment – Number of call attempts

Implementation (2010)  National dual-frame landline and cell phone RDD survey  Separate reports for landline and cell phone samples – Many interviewers worked on both samples, but some could do better on one sample than the other  Started excluding bilingual interviewers – Exhibited substantially different performance

Implementation (2010) continued  The interviewer performance report has multiple goals – Identify lowest performers – Identify improvement of lowest performers  All models estimated twice – Cumulative data – Weekly data only

Graphical Display

Tabular Display

Adoption ( )  Initial period during which supervisors tracked the usual raw paradata-based metrics and the model-based interviewer performance metrics in parallel  Over time, adoption of the model-based metrics along with the addition of reference performance thresholds led to: – consistent identification of interviewers – follow-up after intervention using the same standardized metrics

Experiment ( )  General finding that interviewer training (and feedback) can help lowest performing interviewers (Groves and McGonagle, 2001)  By modeling the paradata to evaluate interviewer performance, could we: – Increase the average interviewer performance? – Reduce the variability in interviewer performance (by more accurately and quickly identifying the lowest performing interviewers)?

Experimental Design  Randomly assign interviewers to a control and an experimental condition – Control: Feedback based on standard (non- modeled paradata) report – Experimental: Use an additional report with a model-based estimate of interviewer performance  Provide feedback to lowest-performing fifth of interviewers identified in each condition

Results – Average Interviewer Performance n.s. (test performed on log-transformed ratio)

Results – Variability in Interviewer Performance across Interviewers n.s.

Summary and Conclusions  For both outcomes (average performance and variability across interviewers) differences were in the expected directions, but not significant (and relatively small)  Results were very similar across both types of samples (landline and cell phone)  We interpret these results as encouraging, providing impetus for further development and investigation

Next Steps  Identifying lower-performing interviewers early in data collection is essential, but not sufficient in addressing performance; similar experimentation is needed to identify more effective feedback and other interventions

Current Phase  Need multiple paradata-based metrics for each interviewer to address: – Efficiency – Nonresponse – Measurement error  Develop and implement other metrics, and augment the interviewer performance report – Refusal rates – Data quality measures (e.g., item nonresponse rates) – Coded interviewer behaviors from monitoring sessions

In-Person Interview Surveys  This model relies on correcting for departures from randomization (the quasi-randomization in centralized call centers)  How well could such paradata-based models correct for lack of any randomization in face to face surveys?  Models have been developed, but not tested: – West and Groves (2013) – Erdman, Adams, and O’Hare (2016)

Thank you Emilia Peytcheva