Developing an evaluation of professional development Webinar #4: Going Deeper into Analyzing Results 1.

Slides:



Advertisements
Similar presentations
Standardized Scales.
Advertisements

Mywish K. Maredia Michigan State University
A Guide to Education Research in the Era of NCLB Brian Jacob University of Michigan December 5, 2007.
How Do We Know if a Charter School is Really Succeeding? – Various Approaches to Investigating School Effectiveness October 2012 Missouri Charter Public.
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Experimental Research Designs
Reading the Dental Literature
Estimation and Reporting of Heterogeneity of Treatment Effects in Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare.
Standards for Qualitative Research in Education
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Introduction to Research
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
Business research methods: data sources
Chapter 3 Preparing and Evaluating a Research Plan Gay and Airasian
Today Concepts underlying inferential statistics
Richard M. Jacobs, OSA, Ph.D.
Studying treatment of suicidal ideation & attempts: Designs, Statistical Analysis, and Methodological Considerations Jill M. Harkavy-Friedman, Ph.D.
Reliability & Validity Qualitative Research Methods.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar May 2014 Note: These slides are intended as guidance.
McGraw-Hill © 2006 The McGraw-Hill Companies, Inc. All rights reserved. The Nature of Research Chapter One.
Qualitative Research.
Selecting a Research Design. Research Design Refers to the outline, plan, or strategy specifying the procedure to be used in answering research questions.
Chapter 1: Introduction to Statistics
RESEARCH A systematic quest for undiscovered truth A way of thinking
Moving from Development to Efficacy & Intervention Fidelity Topics National Center for Special Education Research Grantee Meeting: June 28, 2010.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Chapter 8 Introduction to Hypothesis Testing
Planning an Applied Research Project Chapter 7 – Forms of Quantitative Research © 2014 John Wiley & Sons, Inc. All rights reserved.
CHAPTER III IMPLEMENTATIONANDPROCEDURES.  4-5 pages  Describes in detail how the study was conducted.  For a quantitative project, explain how you.
Chapter 11: Qualitative and Mixed-Method Research Design
Chapter 3 Research Methods Used to Study Child Behavior Disorders.
Systematic Review Module 7: Rating the Quality of Individual Studies Meera Viswanathan, PhD RTI-UNC EPC.
Consumer behavior studies1 CONSUMER BEHAVIOR STUDIES STATISTICAL ISSUES Ralph B. D’Agostino, Sr. Boston University Harvard Clinical Research Institute.
Validity of Qualitative Data Carolyn Seaman UMBC Baltimore, USA Presented to the International Software Engineering Network (ISERN) 21 August 2001 Glasgow,
Eloise Forster, Ed.D. Foundation for Educational Administration (FEA)
Quantitative and Qualitative Approaches
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
Preliminary Results – Not for Citation Strengthening Institutions Program Webinar on Competitive Priority on Evidence April 11, 2012 Note: These slides.
CDIS 5400 Dr Brenda Louw 2010 Validity Issues in Research Design.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar April 25, 2012 Note: These slides are intended as.
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
1 Copyright © 2011 by Saunders, an imprint of Elsevier Inc. Chapter 8 Clarifying Quantitative Research Designs.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
WWC Standards for Regression Discontinuity Study Designs June 2010 Presentation to the IES Research Conference John Deke ● Jill Constantine.
EDCI 696 Dr. D. Brown Presented by: Kim Bassa. Targeted Topics Analysis of dependent variables and different types of data Selecting the appropriate statistic.
The Multiple Comparisons Problem in IES Impact Evaluations: Guidelines and Applications Peter Z. Schochet and John Deke June 2009, IES Research Conference.
SP_IRS Introduction to Research in Special and Inclusive Education(Autumn 2015) Lecture 1: Introduction Lecturer: Mr. S. Kumar.
META-ANALYSIS, RESEARCH SYNTHESES AND SYSTEMATIC REVIEWS © LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON.
One-Way Analysis of Covariance (ANCOVA)
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
SOCW 671: #5 Measurement Levels, Reliability, Validity, & Classic Measurement Theory.
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Securing External Federal Funding Janice F. Almasi, Ph.D. Carol Lee Robertson Endowed Professor of Literacy University of Kentucky
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Chapter 6 Conducting & Reading Research Baumgartner et al Chapter 6 Selection of Research Participants: Sampling Procedures.
Introduction to Research. Purpose of Research Evidence-based practice Validate clinical practice through scientific inquiry Scientific rational must exist.
Developing an evaluation of professional development Webinar #2: Going deeper into planning the design 1.
Characteristics of Studies that might Meet the What Works Clearinghouse Standards: Tips on What to Look For 1.
Preliminary Results – Not for Citation Investing in Innovation (i3) Fund Evidence & Evaluation Webinar 2015 Update Note: These slides are intended as guidance.
Approaches to quantitative data analysis Lara Traeger, PhD Methods in Supportive Oncology Research.
What is Research Design? RD is the general plan of how you will answer your research question(s) The plan should state clearly the following issues: The.
Developing an evaluation of professional development Overview session: Critical elements of professional development planning and evaluation 1.
June 25, Regional Educational Laboratory - Southwest Review of Evidence on the Effects of Teacher Professional Development on Student Achievement:
CRITICALLY APPRAISING EVIDENCE Lisa Broughton, PhD, RN, CCRN.
Chapter 6 Selecting a Design. Research Design The overall approach to the study that details all the major components describing how the research will.
Chapter 12 Quantitative Questions and Procedures.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Writing a sound proposal
Presentation transcript:

Developing an evaluation of professional development Webinar #4: Going Deeper into Analyzing Results 1

Information and materials mentioned or shown during this presentation are provided as resources and examples for the viewer's convenience. Their inclusion is not intended as an endorsement by the Regional Educational Laboratory Southeast or its funding source, the Institute of Education Sciences (Contract ED-IES-12-C- 0011). In addition, the instructional practices and assessments discussed or shown in these presentations are not intended to mandate, direct, or control a State’s, local educational agency’s, or school’s specific instructional content, academic achievement system and assessments, curriculum, or program of instruction. State and local programs may use any instructional content, achievement system and assessments, curriculum, or program of instruction they wish. 2

Webinar 4: Outline Considerations for quantitative analyses – Dr. Sharon Koon – WWC evidence standards & strong studies – Calculating attrition – Calculating baseline equivalence – Statistical adjustments Considerations for qualitative analyses – Dr. La’Tara Osborne-Lampkin Question & answer session 3

CONSIDERATIONS FOR QUANTITATIVE ANALYSES Dr. Sharon Koon 4

Distinction between WWC evidence standards and additional qualities of strong studies WWC design considerations : – Two groups—treatment (T) and comparison (C). – For randomized controlled trials (RCTs), low attrition – For quasi-experimental designs (QEDs), baseline equivalence between T and C groups. – Contrast between T and C groups measures impact of the treatment. – Valid and reliable outcome data used to measure the impact of a treatment. – No known confounding factors. – Outcome(s) not overaligned with the treatment. – Same data collection process—same instruments, same time/year—for the T and C groups. 5 Source: Experiments-in-Education-Version-2.pdf

Distinction between WWC evidence standards and additional qualities of strong studies (cont.) Additional qualities of strong studies: – Pre-specified and clear primary and secondary research questions. – Generalizability of the study results. – Clear criteria for research sample eligibility and matching methods. – Sample size large enough to detect meaningful and statistically significant differences between the T and C groups overall and for specific subgroups of interest. – Analysis methods reflect the research questions, design, and sample selection procedures. – A clear plan to document the implementation experiences of the T and C conditions. 6 Source: Experiments-in-Education-Version-2.pdf

Determinants of a What Works Clearinghouse (WWC) study rating Source: 7

Topics for discussion Attrition Baseline equivalence – Calculation of baseline equivalence – Adjustments for nonequivalence Effect-size corrections – Cluster correction – Multiple comparison correction Handling missing data 8

Attrition For RCTs, the WWC is concerned about both overall attrition (i.e., the rate of attrition for the entire sample) and differential attrition (i.e., the difference in the rates of attrition for the intervention and comparison groups) because both types of attrition contribute to the potential bias of the estimated effect. Source: 9

Attrition (cont.) Overall attrition = Number without observed data/number randomized Differential attrition = [T without observed data/number T randomized] – [C without observed data/number C randomized] Attrition boundaries: liberal or conservative In order to be deemed an RCT with low attrition, a cluster RCT that reports an individual-level analysis must have low attrition at two levels. First, it must have low attrition at the cluster level. Second, the study must have low attrition at the subcluster level, with attrition based only on the clusters remaining in the sample. Source: 10

Attrition (cont.) Source: 11

Baseline equivalence For continuous outcomes, it is determined by the difference between the mean outcome for the T group and the mean outcome for the C group, divided by the pooled within-group standard deviation of the outcome measure (i.e., standardized mean difference). For dichotomous outcomes, it is determined by the difference in the probability of the occurrence of an event. Source: 12

Statistical adjustment for nonequivalence at baseline For differences in baseline characteristics that are between 0.05 and 0.25 standard deviations, the analysis must include a statistical adjustment for the baseline characteristics to meet the baseline equivalence requirement. A number of different techniques can be used, including regression adjustment and analysis of covariance (ANCOVA). The critical factor is that the appropriate baseline characteristics must be included in the analysis at the individual level (i.e., the unit of analysis). Source: 13

Difference-in-difference adjustment The WWC applies this adjustment to effect size calculations based on unadjusted group means when the study is: – a QED with differences in baseline characteristics less than.05 – an RCT with low attrition and differences in baseline characteristics – an RCT with high attrition and differences in baseline characteristics less than.05 Source: 14

Cluster correction A “mismatch” problem occurs when random assignment is carried out at the cluster level (e.g., school level) and the analysis is conducted at the individual level (e.g., teacher level), but the correlation among students within the same clusters is ignored in computing the standard errors of the impact estimates. The standard errors of the impact estimates generally will be underestimated, thereby leading to overestimates of statistical significance. Source: 15

Cluster correction (cont.) The WWC computes clustering-corrected statistical significance estimates. The basic approach to the clustering correction is first to compute the t-statistic corresponding to the effect size that ignores clustering and then correct both the t-statistic and the associated degrees of freedom for clustering based on sample sizes, number of clusters, and an estimate of the intra-class correlation (ICC). The default ICC value is 0.20 for achievement outcomes and 0.10 for behavioral and attitudinal outcomes. The statistical significance estimate corrected for clustering is then obtained from the t-distribution using the corrected t- statistic and degrees of freedom. Source: 16

Multiple comparisons correction Repeated tests of highly correlated outcomes will lead to a greater likelihood of mistakenly concluding that the differences in means for outcomes of interests between the T and C groups are significantly different from zero (called Type I error in hypothesis testing). The WWC uses the Benjamini-Hochberg (BH) correction to reduce the possibility of making this type of error. Source: 17

Multiple comparisons correction (cont.) The BH correction is used in three types of situations: studies that estimated effects of the intervention for multiple outcome measures in the same outcome domain using a single comparison group, studies that estimated effects of the intervention for a given outcome measure using multiple comparison groups, and studies that estimated effects of the intervention for multiple outcome measures in the same outcome domain using multiple comparison groups. Source: 18

Handling missing data RCTs & QEDs: Complete case analysis with no adjustment for missing data (for example, unadjusted means) or with covariate adjustment Low-attrition RCTs only: – Complete case analysis with nonresponse weights – Multiple imputation, separately by condition and using an established method (however, can not be used to meet the attrition standard) – Maximum likelihood, separately by condition and using an established method Source: 19

CONSIDERATIONS FOR QUALITATIVE ANALYSES Dr. La’Tara Osborne-Lampkin 20

(Minchiello et al., 1990, p.5) QualitativeQuantitative ConceptualConcerned with understanding human behavior from the informant’s perspective Concerned with discovering facts about human behavior Assumes a dynamic & negotiated reality Assumes a fixed & measurable reality MethodologicalData are collected through participant observation & interviews Data are collected through measuring things Data are analyzed by themes from descriptions by informants Data are analyzed through numerical comparisons & statistical inferences Data are reported through narratives and the language of the informant Data are reported through statistical analyses 21

When analyzing qualitative data  Use an iterative approach  guided by prior data collection and analysis  Use multiple, “reliable” researchers for analysis and interpretation  Document and outline steps and decisions made for data analysis (i.e., develop an audit trail)  Document the basis for inferences  Establish structural corroboration or coherence (Miles & Huberman, 1994; Miles, Huberman & Soldana, 2014, Patton, 1987) 22

Five key considerations Triangulate data Ensure representativeness Look for competing explanations Analyze negative cases Keep methods & data in context 23

Triangulate data  Use multiple methods to study a program  Collect multiple types of data on the same question  Use different interviewers to avoid biases of any one different data collector and interviewers working alone  Use multiple perspectives (or theories) to interpret a set of data 24

Practical strategies to triangulate Data sources  Compare interview data  focus group  semi-structured interview data  With  observational data  documents  Validate observational data with documents Participants  Compare what participants say in public with what they say in private (for example, focus group data vs. individual interview data)  Check consistency of what participants do and say over time  Compare the perspectives of individuals within and across stakeholder groups 25

Verify “fit” and “work” “representativeness” in the data… 26

Look for rival and competing explanations May employ an inductive or logical process Use data to support alternative explanations that are grounded in logic and theory Weight the evidence and look for best fit 27

Negative cases “Exceptions that prove the rule” Counter evidence 28

Keeping methods and data in context Limit conclusions to: those situations, time periods, persons, and contexts for which the data are applicable. Keeping things in context is the cardinal principle of qualitative analysis (Patton, 1987). 29

Questions & Answers Homework: Bring remaining questions to session 5 30

Developing an evaluation of professional development Webinar 5: Going Deeper into Interpreting Results & Presenting Findings 1/21/2016, 2:00pm 31