Meta-Analysis: A Gentle Introduction to Research Synthesis Gianna Rendina-Gobioff Jeff Kromrey Research Methods in a Nutshell College of Education Presentation.

Slides:



Advertisements
Similar presentations
Conducting a Meta-Analysis Jose Ramos Southern Methodist University Paper presented at the annual meeting of the Southwest Educational Research Association,
Advertisements

What is meta-analysis? ESRC Research Methods Festival Oxford 8 th July, 2010 Professor Steven Higgins Durham University
Significance and effect sizes What is the problem with just using p-levels to determine whether one variable has an effect on another? Don’t EVER just.
Funded through the ESRC’s Researcher Development Initiative Department of Education, University of Oxford Session 2.3 – Publication bias.
The Campbell Collaborationwww.campbellcollaboration.org Moderator analyses: Categorical models and Meta-regression Terri Pigott, C2 Methods Editor & co-Chair.
Introduction to Meta-Analysis Joseph Stevens, Ph.D., University of Oregon (541) , © Stevens 2006.
Today Concepts underlying inferential statistics
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Meta-Analysis. Why Meta-Analysis? There is an urgent need for reliable summaries of primary research in music therapy. 1. Music therapists can not keep.
CORRELATIO NAL RESEARCH METHOD. The researcher wanted to determine if there is a significant relationship between the nursing personnel characteristics.
Part 3 of 3 By: Danielle Davidov, PhD & Steve Davis, MSW, MPA INTRODUCTION TO RESEARCH: SAMPLING & DESIGN.
S-005 Types of research in education. Types of research A wide variety of approaches: –Theoretical studies –Summaries of studies Reviews of the literature.
September 26, 2012 DATA EVALUATION AND ANALYSIS IN SYSTEMATIC REVIEW.
Are the results valid? Was the validity of the included studies appraised?
9.0 A taste of the Importance of Effect Size The Basics of Effect Size Extraction and Statistical Applications for Meta- Analysis Robert M. Bernard Philip.
Copyright © 2008 by Pearson Education, Inc. Upper Saddle River, New Jersey All rights reserved. John W. Creswell Educational Research: Planning,
Funded through the ESRC’s Researcher Development Initiative
Advanced Statistics for Researchers Meta-analysis and Systematic Review Avoiding bias in literature review and calculating effect sizes Dr. Chris Rakes.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
The Effect of Computers on Student Writing: A Meta-Analysis of Studies from 1992 to 2002 Amie Goldberg, Michael Russell, & Abigail Cook Technology and.
Education Research 250:205 Writing Chapter 3. Objectives Subjects Instrumentation Procedures Experimental Design Statistical Analysis  Displaying data.
L 1 Chapter 12 Correlational Designs EDUC 640 Dr. William M. Bauer.
Analyzing and Interpreting Quantitative Data
Quantitative Research. Overview Non-experimental QualitativeCase study Phenomenology Ethnography Historical Literature Review QuantitativeObservational.
Statistical Applications for Meta-Analysis Robert M. Bernard Centre for the Study of Learning and Performance and CanKnow Concordia University December.
Funded through the ESRC’s Researcher Development Initiative Prof. Herb MarshMs. Alison O’MaraDr. Lars-Erik Malmberg Department of Education, University.
Assumes that events are governed by some lawful order
Meta-analysis and “statistical aggregation” Dave Thompson Dept. of Biostatistics and Epidemiology College of Public Health, OUHSC Learning to Practice.
Meta-analysis 統合分析 蔡崇弘. EBM ( evidence based medicine) Ask Acquire Appraising Apply Audit.
The Campbell Collaborationwww.campbellcollaboration.org C2 Training: May 9 – 10, 2011 Introduction to meta-analysis.
Wim Van den Noortgate Katholieke Universiteit Leuven, Belgium Belgian Campbell Group Workshop systematic reviews.
Statistical Models for the Analysis of Single-Case Intervention Data Introduction to:  Regression Models  Multilevel Models.
EDCI 696 Dr. D. Brown Presented by: Kim Bassa. Targeted Topics Analysis of dependent variables and different types of data Selecting the appropriate statistic.
Developing a Review Protocol. 1. Title Registration 2. Protocol 3. Complete Review Components of the C2 Review Process.
META-ANALYSIS, RESEARCH SYNTHESES AND SYSTEMATIC REVIEWS © LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON.
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
One-Way Analysis of Covariance (ANCOVA)
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
© (2015, 2012, 2008) by Pearson Education, Inc. All Rights Reserved Chapter 11: Correlational Designs Educational Research: Planning, Conducting, and Evaluating.
Chapter 6: Analyzing and Interpreting Quantitative Data
Lecture 5 EPSY 642 Victor Willson Fall EFFECT SIZE DISTRIBUTION Hypothesis: All effects come from the same distribution What does this look like.
Sifting through the evidence Sarah Fradsham. Types of Evidence Primary Literature Observational studies Case Report Case Series Case Control Study Cohort.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
What is Research? research is an unusually stubborn and persisting effort to think straight which involves the gathering and the intelligent use of relevant.
EBM --- Journal Reading Presenter :呂宥達 Date : 2005/10/27.
Systematic Synthesis of the Literature: Introduction to Meta-analysis Linda N. Meurer, MD, MPH Department of Family and Community Medicine.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Chapter Eight: Quantitative Methods
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Introduction to Validity True Experiment – searching for causality What effect does the I.V. have on the D.V. Correlation Design – searching for an association.
Handout Twelve: Design & Analysis of Covariance
Handout Six: Sample Size, Effect Size, Power, and Assumptions of ANOVA EPSE 592 Experimental Designs and Analysis in Educational Research Instructor: Dr.
Funded through the ESRC’s Researcher Development Initiative Department of Education, University of Oxford Session 2.1 – Revision of Day 1.
Systematic Reviews and Meta-analyses. Introduction A systematic review (also called an overview) attempts to summarize the scientific evidence related.
Chapter 14 Research Synthesis (Meta-Analysis). Chapter Outline Using meta-analysis to synthesize research Tutorial example of meta-analysis.
How Psychologists Do Research Chapter 2. How Psychologists Do Research What makes psychological research scientific? Research Methods Descriptive studies.
Wim Van den Noortgate Katholieke Universiteit Leuven, Belgium Belgian Campbell Group Workshop systematic reviews.
An Application of Multilevel Modelling to Meta-Analysis, and Comparison with Traditional Approaches Alison O’Mara & Herb Marsh Department of Education,
Appropriate use of Design Effects and Sample Weights in Complex Health Survey Data: A Review of Articles Published using Data from Add Health, MTF, and.
Reviewing systematic reviews: meta- analysis of What Works Clearinghouse computer-assisted interventions. November 2011 American Evaluation Association.
Chapter 22 Inferential Data Analysis: Part 2 PowerPoint presentation developed by: Jennifer L. Bellamy & Sarah E. Bledsoe.
Week Seven.  The systematic and rigorous integration and synthesis of evidence is a cornerstone of EBP  Impossible to develop “best practice” guidelines,
Statistics & Evidence-Based Practice
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Selecting the Best Measure for Your Study
Supplementary Table 1. PRISMA checklist
Statistical Models for the Analysis of Single-Case Intervention Data
Chapter Eight: Quantitative Methods
Meta-analysis, systematic reviews and research syntheses
Presentation transcript:

Meta-Analysis: A Gentle Introduction to Research Synthesis Gianna Rendina-Gobioff Jeff Kromrey Research Methods in a Nutshell College of Education Presentation December 8, 2006

Discussion Outline Overview Types of research questions Literature search and retrieval Coding and dependability Effect sizes Describing results Testing hypotheses Threats to validity Reporting meta-analyses References worth pursuing

Overview Summarization of empirical studies using quantitative methods Results Estimated weighted mean effect size Confidence interval around mean effect size (or test null hypothesis about mean effect size) Homogeneity of effect sizes Tests of moderators

Overview: Why Meta-Analyze? Strength in numbers Several ‘non-significant’ differences may be significant when combined Strength in diversity Generalizability across variety of participants, settings, instruments Identification of moderating variables Good way to look at the forest rather than the trees What do we think we know about a phenomenon? How well do we know it? What remains to be investigated? It’s fun!

Overview: Stages of Meta-Analysis Formulate problem Draw sample / collect observations Measure observations Analyze data Interpret data Disseminate

Types of Research Questions: Treatments Is the treatment (in general) effective? How effective? Does treatment effectiveness vary by Participant characteristics? Treatment characteristics? Research method characteristics? Is the treatment ineffective in some conditions?

Types of Research Questions: Relationships What is the relationship (in general)? Direction? Strength? Does direction or strength of relationship vary by Participant characteristics? Treatment characteristics? Research method characteristics? Is the relationship not evident in some conditions?

Literature Search and Retrieval Decisions to make before searching the literature Inclusion/Exclusion criteria for sources Types of publication  Language and country of publication  Dissemination: Journal, presentation, unpublished Study characteristics  Participant characteristics  Information reported  Timeframe  Type of design  Measures

Literature Search and Retrieval Decisions to make before searching the literature Search strategies Keywords Databases  ERIC, PsychInfo, GoogleScholar, Web of Science Other  Key researchers  Listservs  Websites  Reference sections of articles

Coding of Studies Record Study inclusion/exclusion characteristics Effect size(s) Multiple measures? Subsamples? Different times? Other relevant variables Research design (sampling, controls, treatment, duration) Participant attributes (age, sex, race/ethnicity, inclusion/exclusion) Settings (geography, classrooms, laboratory) Dissemination characteristics (journal, conference, dissertation, year, Dr. B)

Coding of Studies (Cont’d) Written codebook and coding forms Goldilock’s principle: not too coarse, not too fine. Training and calibration of coders Beware of drift Estimating reliability of coders

Study Coding Form Meta-Analysis Coding Part I: Increased levels of stress will reduce the likelihood of ART treatment success. STUDY TITLE: I. Qualifying the study: Answer the following questions as either “yes” or “no”. Does the study involve women participating in an ART treatment program? Does the study focus on the relationship between stress and ART treatment outcomes? Was the study conducted between January 1985 and December 2003? Does the study employ a prospective design? Does the study report outcome measures of stress or anxiety as well as ART treatment outcomes? If the answer to each of the above questions is yes, the study qualifies for inclusion in the meta-analysis.

II.Coding the study: A. Publication Characteristics 1. Title of the study: 2. Year of Publication: 3. Authors: B.Ecological Characteristics 1. Age of Female Participants: Mean: Range: 2. Country: 3. Race: WhiteN:%: BlackN:%: HispanicN:%: Asian / Pacific IslanderN:%: American IndianN:%: OtherN:%:

Duration of Psychoeducational Intervention (Please choose) a. Daily for duration of ART treatment b. 1 – 3 sessions during ART treatment c. 6 weeks during ART treatment d. 8 weeks during ART treatment e. 10 weeks during ART treatment f. Other: Length of Psychoeducational Intervention (Please choose) a. 1 hour b. 1.5 hours c. 2 hours d. Other: Frequency of Psychoeducational Intervention (Please choose) a. Daily b. Weekly c. Bi-Weekly d. Other:

Effect Size How false is the null hypothesis? How effective is the treatment? How strong is the relationship? Independent of sample size (more or less) Useful in primary studies and in meta- analysis Links to power Descriptive statistic (big enough to care?)

Effect Size (Cont’d) Jacob Cohen Statistical Power Analysis for the Behavioral Sciences Anytime a statistical hypothesis is tested, an effect size is lurking in there somewhere Small, medium, large effects Medium effect size is big enough to be seen by the naked eye of the careful but naïve observer

Effect Size: Standardized Mean Difference Population effect size Sample effect size Small=.20, Medium=.50, Large=.80

Effect Size: Chi-square Tests Population effect size Sample effect size Small=.10, Medium=.30, Large=.50

Effect Size: ANOVA and Regression ANOVA Regression (test of R 2 ) Regression (test of R 2 change)

Effect Size: Correlation Pearson Product Moment Correlation is an effect size Commonly transformed to z for aggregation and analyses Small=.10, Medium=.30, Large=.50

Knowing Effect Size: Computing from Reported Statistics Article Information:

t(54) = 4.52, p <.05 Effect Size: Computing from Reported Statistics Article Information:

Describing Results: Graphical Displays

Stem and Leaf Plot

Describing Results: Graphical Displays

All observed effect sizes from a single population d1d1 d2d2 d6d6 d5d5 d4d4 d3d3  Testing Hypotheses

Observed effect sizes from two populations d1d1 d2d2 d6d6 d5d5 d4d4 d3d3   Males Females Testing Hypotheses

Testing Hypotheses: Fixed Effects vs. Random Effects

Fixed Effects Assumes one population effect size Effect size variance = sampling error (subjects) Weights represent study variance due to sampling error associated with the subjects (sample size) Testing Hypotheses: Fixed Effects vs. Random Effects

Random Effects Assumes population effect size is a normal distribution of values (i.e. not one effect size) Effect size variance = sampling error (subjects) + random effects (study) Weights represent study variance due to sampling error associated with the subjects (sample size) and sampling of studies (random effects variance component) Testing Hypotheses: Fixed Effects vs. Random Effects

Fixed Effects vs. Random Effects: Which model to use? Aspects to consider: Statistics – decision based on the outcome of the homogeneity of effect sizes statistic (conditionally random-effects) Desired Inferences – decision based on the inferences that the researcher would like to make  Conditional Inferences (fixed effect model): Researcher can only generalize to the studies included in the meta-analysis  Unconditional Inferences (random effect model): Researcher can generalize beyond the studies included in the meta-analysis Number of studies – when the number of studies is small fixed effects may be more appropriate Testing Hypotheses: Fixed Effects vs. Random Effects

Fixed effects weight Random effects weight For standardized mean difference: Testing Hypotheses: Estimation of Weights

Fixed effects Random effects Testing Hypotheses: Weighted Mean Effect Size

Also called Random Effects Variance Component (REVC), symbolized with Used to calculate random effects weights Three methods to calculate Observed variance Q based Maximum likelihood Testing Hypotheses: Estimates of Effect Size Variance

Observed variance: Q based: Maximum likelihood: Testing Hypotheses: Estimates of Effect Size Variance

Significance testing: Confidence interval (95% CI): Testing Hypotheses: Significance Testing and Confidence Intervals (CI)

Focused test of between group differences General test of homogeneity of effect sizes Testing Hypotheses: Mean and Individual Effect Size Differences

Generalization of the Q test Continuous or categorical moderators X i are potential moderating variables Test for moderating effect Testing Hypotheses: Meta-Analytic Regression Model

Threats to Validity Sources Primary studies – unreliability, restriction of range, missing effect sizes (publication bias), incompatible constructs, and poor quality Meta-analysis processes – incomplete data collection (publication bias), inaccurate data collection, poor methodology, and inadequate power

Threats to Validity Apples and Oranges Dependent Effect Sizes File Drawer/Publication Bias Methodological Rigor Power

Threats to Validity Apples and Oranges Are the studies being analyzed similar regarding: Constructs examined Measures Participants (sampled from same population?) Analyses Dependent Effect Sizes Participants cannot contribute to the mean effect size more than once

Threats to Validity: Publication Bias Publication Bias = Studies unavailable to the meta-analyst due to lack of publication acceptance or submission (termed “file drawer problem” by Rosenthal, 1979) Pattern in the literature

Threats to Validity: Publication Bias Publication Bias Detection Methods Visual interpretation Funnel plot display Statistical methods Begg Rank Correlation (variance or sample size) Egger Regression Funnel Plot Regression Trim and Fill

Threats to Validity Methodological Rigor of Primary Studies Set criteria for inclusion Include various levels of rigor; then code and use in meta-analytic analyses (moderators or quality weights) Power Enough studies collected to achieve significant findings?

Reporting Meta-Analyses: Pertinent Information to Include Details regarding the search criteria and retrieval Coding process including rater reliability Describe effect sizes graphically Analyses Mean effect size (significance test and CI) Fixed vs. Random Effects model Homogeneity of effect sizes Tests for moderators How threats to validity were addressed

For Further Reading & Thinking Bangert-Drowns, R.L. (1986). Review of developments in meta-analysis method. Psychological Bulletin, 99, Cohen, J. (1992). A power primer. Psychological Bulletin, 112, Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). New York: Academic Press. Cooper H. & Hedges, L. (1994). The handbook of research synthesis. New York: Russell Sage Foundation. Fern, E. F. & Monroe, K. B. (1996). Effect size estimates: Issues and problems in interpretation. Journal of Consumer Research, 23, Grissom R.J. & Kim J.J. (2001). Review of assumptions and problems in the appropriate conceptualization of effect size. Psychological Methods, 6(2), p Hedges, L.V. & Olkin, I. (1985). Statistical methods for meta-analysis. San Diego, CA: Academic Press. Hedges, L.V. & Vevea, J. (1998). Fixed- and random-effects models in meta-analysis. Psychological Methods, 3, Hedges, L. V., & T. D. Pigott The power of statistical tests in meta-analysis. Psychological Methods, 6, 203–17. Hedges, L. V., & T. D. Pigott The power of statistical tests for moderators in meta-analysis. Psychological Methods, 9, 426–45.

For Further Reading & Thinking Hogarty, K. Y. & Kromrey, J. D. (2000). Robust effect size estimates and meta-analytic tests of homogeneity. Proceedings of SAS Users’ Group International, Hogarty, K. Y. & Kromrey, J. D. (2001, April). We’ve been reporting some effect sizes: Can you guess what they mean? Paper presented at the annual meeting of the American Educational Research Association, Seattle. Hogarty, K. Y. & Kromrey, J. D. (2003). Permutation tests for linear models in meta-analysis: Robustness and power under non-normality and variance heterogeneity. Proceedings of the American Statistical Association. Alexandria, VA: American Statistical Association. Huberty, C. J. & Lowman, L. L. (2000). Group overlap as a basis for effect size. Educational and Psychological Measurement, 60, 543 – 563. Hunter, J. E., & Schmidt, F. L. (2004). Methods of meta-analysis: Correcting error and bias in research findings (2nd edition). Newbury Park, CA: Sage. Hillsdale, NJ. Kromrey, J. D., Ferron, J. D., Hess, M. R., Hogarty, K. Y. & Hines, C. V. (2005, April). Robust Inference in Meta-analysis: Comparing Point and Interval Estimates Using Standardized Mean Differences and Cliff’s Delta. Annual meeting of the American Educational Research Association, Montreal. Kromrey, J. D. & Foster ‑ Johnson, L. (1996). Determining the efficacy of intervention: The use of effect sizes for data analysis in single ‑ subject research. Journal of Experimental Education, 65, 73 ‑ 93.

For Further Reading & Thinking Kromrey, J. D. & Hogarty, K. Y. (2002). Estimates of variance components in random effects meta- analysis: Sensitivity to violations of normality and variance homogeneity. Proceedings of the American Statistical Association. Alexandria, VA: American Statistical Association. Kromrey, J. D., Hogarty, K. Y., Ferron, J. M., Hines, C. V. & Hess, M. R. (2005, August). Robustness in Meta-Analysis: An Empirical Comparison of Point and Interval Estimates of Standardized Mean Differences and Cliff’s Delta. Proceedings of the American Statistical Association Joint Statistical Meetings. Kromrey, J. D. & Foster-Johnson, L. (1999, February). Effect sizes, cause sizes and the interpretation of research results: Confounding effects of score variance on effect size estimates. Paper presented at the annual meeting of the Eastern Educational Research Association, Hilton Head, South Carolina. Kromrey, J. D. & Rendina-Gobioff, G. (2006). On knowing what we don't know: An empirical comparison of methods to detect publication bias in meta-analysis. Educational and Psychological Measurement, 66, Lipsey, M. W., & Wilson, D. B. (1993). The efficacy of psychological, educational, and behavioral treatment: Confirmation from meta-analysis. American Psychologist, 48, 1181–1209.

For Further Reading & Thinking Lipsey, M. W., & Wilson, D. B. (2001). Practical Meta-analysis. Thousand Oaks: Sage. National Research Council (1992). Combining information: Statistical issues and opportunities for research. Washington, DC: National Academy of Science Press. Rendina-Gobioff, G. (2006). Detecting Publication Bias in Random Effects Meta-Analysis: An Empirical Comparison of Statistical Methods. Unpublished doctoral dissertation, University of South Florida, Tampa. Rendina-Gobioff, G., Kromrey, J. D., Dedrick, R. F., & Ferron, J. M. (2006, November). Detecting Publication Bias in Random Effects Meta-Analysis: An Investigation of the Performance of Statistical Methods. Paper presented at the annual meeting of the Florida Educational Research Association, Jacksonville. Rendina-Gobioff, G. & Kromrey, J. D. (2006, October). PUB_BIAS: A SAS® Macro for Detecting Publication Bias in Meta-Analysis. Paper presented at the annual meeting of the Southeast SAS Users Group, Atlanta Rosenthal, R. (1995). Writing meta-analytic reviews. Psychological Bulletin, 118, Sutton, A. J., Abrams, K. R., Jones, D. R., Sheldon, T. A., & Song, F. (2000). Methods of meta-analysis in medical research. New York: Wiley. Van den Noortgate, W., & Onghena, P. (2003). Multilevel meta-analysis: A comparison with traditional meta-analytical procedures. Educational and Psychological Measurement, 63,

Thank You Now, let’s just talk…