Longitudinal Experiments Larry V. Hedges Northwestern University Prepared for the IES Summer Research Training Institute July 28, 2010.

Slides:



Advertisements
Similar presentations
Multiple Regression Analysis
Advertisements

The Multiple Regression Model.
Statistical Analysis Overview I Session 2 Peg Burchinal Frank Porter Graham Child Development Institute, University of North Carolina-Chapel Hill.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Random Assignment Experiments
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Sampling Distributions and Sample Proportions
Variance reduction techniques. 2 Introduction Simulation models should be coded such that they are efficient. Efficiency in terms of programming ensures.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Environmental Data Analysis with MatLab Lecture 23: Hypothesis Testing continued; F-Tests.
ANOVA: ANalysis Of VAriance. In the general linear model x = μ + σ 2 (Age) + σ 2 (Genotype) + σ 2 (Measurement) + σ 2 (Condition) + σ 2 (ε) Each of the.
Experimental Design, Statistical Analysis CSCI 4800/6800 University of Georgia Spring 2007 Eileen Kraemer.
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 4. Further Issues.
Chapter 4 Multiple Regression.
Clustered or Multilevel Data
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 4. Further Issues.
Ratio estimation with stratified samples Consider the agriculture stratified sample. In addition to the data of 1992, we also have data of Suppose.
Introduction to Regression Analysis, Chapter 13,
Analysis of Clustered and Longitudinal Data
Correlation and Regression
Objectives of Multiple Regression
Introduction to Multilevel Modeling Using SPSS
Issues in Experimental Design Reliability and ‘Error’
Chapter 5 Sampling Distributions
The Campbell Collaborationwww.campbellcollaboration.org Introduction to Robust Standard Errors Emily E. Tanner-Smith Associate Editor, Methods Coordinating.
CORRELATION & REGRESSION
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Sampling and Nested Data in Practice- Based Research Stephen Zyzanski, PhD Department of Family Medicine Case Western Reserve University School of Medicine.
Discriminant Function Analysis Basics Psy524 Andrew Ainsworth.
Analysis of Variance ( ANOVA )
PARAMETRIC STATISTICAL INFERENCE
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Problems with the Design and Implementation of Randomized Experiments By Larry V. Hedges Northwestern University Presented at the 2009 IES Research Conference.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Testing Hypotheses about Differences among Several Means.
Repeated Measurements Analysis. Repeated Measures Analysis of Variance Situations in which biologists would make repeated measurements on same individual.
Introduction to Multilevel Modeling Stephen R. Porter Associate Professor Dept. of Educational Leadership and Policy Studies Iowa State University Lagomarcino.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Research Methods for Counselors COUN 597 University of Saint Joseph Class # 4 Copyright © 2015 by R. Halstead. All rights reserved.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Adjusted from slides attributed to Andrew Ainsworth
June 11, 2008Stat Lecture 10 - Review1 Midterm review Chapters 1-5 Statistics Lecture 10.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
1 Overview of Experimental Design. 2 3 Examples of Experimental Designs.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Using Specialized Research Designs.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16.
FIXED AND RANDOM EFFECTS IN HLM. Fixed effects produce constant impact on DV. Random effects produce variable impact on DV. F IXED VS RANDOM EFFECTS.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 11: Models Marshall University Genomics Core Facility.
Sampling and Nested Data in Practice-Based Research Stephen Zyzanski, PhD Department of Family Medicine Case Western Reserve University School of Medicine.
Chapter 5 Multilevel Models
Analysis of Experiments
The Mixed Effects Model - Introduction In many situations, one of the factors of interest will have its levels chosen because they are of specific interest.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Chapter 7: The Distribution of Sample Means
Two-Factor Study with Random Effects In some experiments the levels of both factors A & B are chosen at random from a larger set of possible factor levels.
The simple linear regression model and parameter estimation
Analysis for Designs with Assignment of Both Clusters and Individuals
Hypothesis Testing and Confidence Intervals (Part 1): Using the Standard Normal Lecture 8 Justin Kern October 10 and 12, 2017.
Multiple Regression Analysis
CHAPTER 29: Multiple Regression*
Chapter 5 Sampling Distributions
Fixed, Random and Mixed effects
Longitudinal Experiments
An Introductory Tutorial
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Longitudinal Experiments Larry V. Hedges Northwestern University Prepared for the IES Summer Research Training Institute July 28, 2010

What Are Longitudinal Experiments? Longitudinal experiments are experiments with repeated measurements of an outcome on the same people Examples Experiments with immediate and delayed posttests Experiments that track individuals over many performance periods (e.g., school years) Experiments that intend to impact growth rate Experiments that make repeated measurements of behavior (e.g., teacher behavior) to increase precision of measurement

Why Do Longitudinal Experiments? Three reasons for doing longitudinal experiments 1.More than one discrete endpoint is of interest (e.g., immediate vs delayed outcome) 2.Several measures of the outcome are necessary to increase precision or reduce variation (e.g., teacher behavior is averaged over many occasions) 3.The time course of treatment effect (growth trajectory) is of interest (e. g., an intervention is intended to increase the rate of vocabulary acquisition in preschool children) In all three cases, linear combinations of outcomes may be of interest

Why Do Longitudinal Experiments? Unless different outcomes are being compared, there is no need to use longitudinal methods! But, if different outcomes are being compared, outcomes are not independent Thus, longitudinal methods must be used

Modeling Longitudinal Experiments We can describe models for longitudinal experiments via ANOVA or HLM notation We can analyze longitudinal experiments via either ANOVA or HLM There are big advantages to using HLM notation for these models There are even bigger advantages to using HLM for analyses of these models Hence we will primarily use HLM notation in our discussion of longitudinal experiments

Discrete Endpoints The design will typically have at least three levels Measures are nested (clustered) within individuals, individuals are nested (clustered) within schools Level 1 (measures within individuals) Level 2 (individuals within schools) Level 3 (schools) Let Y ijk, the observation on the k th measure for the j th person in the i th school

Discrete Endpoints, Schools Assigned (No Covariates) Level 1 (measure level) Y ijk = β 0ij + ε ijk ε ~ N(0, σ W 2 ) Level 2 (individual level) β 0ij = γ 00i + η 0ij η ~ N(0, σ I 2 ) Level 3 (school level) γ 00i = π 00 + π 01 T i + ξ 0i ξ ~ N(0, σ S 2 ) Where we code the (centered) treatment T j = ½ or - ½, so that π 01 is the ANOVA treatment effect

Discrete Endpoints (Unconditional Model) Note that the ε ijk ’ s are not just measurement errors but also contain differences between outcomes for each individual Similarly the η 0ij ‘s are between individual differences in these quantities Then the ξ 0i ‘s are between-school differences on these quantities That makes the unconditional model difficult to interpret

Discrete Endpoints, Schools Assigned (Comparing Early and Delayed Outcome) Level 1 (measure level) Y ijk = β 0ij + β 1ij D ijk + ε ijk ε ~ N(0, σ W 2 ) Level 2 (individual level) β 0ij = γ 00i + η 0ij η ~ N(0, Σ I ) β 1ij = γ 10i + η 1ij Level 3 (school level) γ 00i = π 00 + π 01 T i + ξ 0i ξ ~ N(0, Σ S ) γ 10i = π 10 + π 11 T i + ξ 1i Note that the η 0ij ’s and η 1ij ’s can be correlated as can the ξ 0i ’s and ξ 1i ’s

Discrete Endpoints, Individuals Assigned (Comparing Early and Delayed Outcome) Level 1 (measure level) Y ijk = β 0ij + β 1ij D ijk + ε ijk ε ~ N(0, σ W 2 ) Level 2 (individual level) β 0ij = γ 00i + γ 01i T i + η 0ij η ~ N(0, Σ I ) β 1ij = γ 10i + γ 11i T i + η 1ij Level 3 (school level) γ 00i = π ξ 00i γ 01i = π ξ 01i ξ ~ N(0, Σ S ) γ 10i = π ξ 10i γ 11i = π ξ 11i Note that the η 0ij ’s and η 1ij ’s can be correlated as can the ξ ’‘s and ξ ’s

Discrete Endpoints (Comparing Early and Delayed Outcome) Note that, in this model, the ε ijk ’ s can be interpreted as measurement errors Similarly the η 0ij ‘s are between individual differences in these quantities and the intraclass correlation ρ I = σ I 2 /(σ s 2 + σ I 2 + σ W 2 ) is a true (individual level) reliability coefficient Then the ξ 0i ‘s are between-school differences on these quantities and the intraclass correlation ρ S = σ S 2 /(σ s 2 + σ I 2 + σ W 2 ) is a true (school level) reliability coefficient

Discrete Endpoints, Schools Assigned (Comparing Early and Delayed Outcome) Covariates can be added at any level of the design But remember that covariates must be variables that cannot have been impacted by treatment assignment Thus time varying covariates (at level 1) are particularly suspect since they may be measured after treatment assignment

Average of Several Measures The design will typically have at least three levels Measures are nested (clustered) within individuals, individuals are nested (clustered) within schools Level 1 (measures within individuals) Level 2 (individuals within schools) Level 3 (schools) Let Y ijk, the observation on the k th measure for the j th person in the i th school with p measures per individual

Average of Several Measures (Treatment Assigned at the School Level) Level 1 (measure level) Y ijk = β 0ij + ε ijk ε ~ N(0, σ W 2 ) Level 2 (individual level) β 0ij = γ 00i + η 0ij η ~ N(0, σ I 2 ) Level 3 (school level) γ 0i = π 00 + π 01 T i + ξ 0i ξ ~ N(0, σ S 2 ) Where we code the (centered) treatment T j = ½ or - ½, so that π 01 is the treatment effect

Average of Several Measures (Treatment Assigned at the Individual Level) Level 1 (measure level) Y ijk = β 0ij + ε ijk ε ~ N(0, σ W 2 ) Level 2 (individual level) β 0ij = γ 00i + γ 01i T ij + η 0ij η ~ N(0, σ I 2 ) Level 3 (school level) γ 00i = π 00 + ξ 0i ξ ~ N(0, Σ S ) γ 01i = π 01 + ξ 1i Where we code the (centered) treatment T j = ½ or - ½, so that π 01 is the treatment effect

Average of Several Measures Note that, in this model, the ε ijk ’ s can be interpreted as like (item level) measurement errors Then the β 0ij ‘s can be interpreted as individual level measures (for the j th person in the i th school) Thus the η 0ij ‘s are between individual differences in these quantities and the quantity ρ I = σ I 2 /(σ s 2 + σ I 2 + σ W 2 /p) is a true (individual level) reliability coefficient Then the ξ 0i ‘s are between-school differences on these quantities and the quantity ρ S = σ S 2 /(σ s 2 + σ I 2 + σ W 2 /p) is a true (school level) reliability coefficient

Growth Trajectories The problem of fitting growth trajectories is more complicated It requires choosing a form for the growth trajectories It also requires choosing a form for the model of individual differences in these growth trajectories Many forms are possible, but polynomials are conventional for two reasons: Any smooth function is approximately a polynomial (Taylor’s Theorem) Polynomials are simple

What is a Polynomial Model? Y ijk = β 0ij + β 1ij t ijk + β 2ij t ijk 2 + β 3ij t ijk 3 + ε ijk Typically, t ijk is a measure of time at the measurement for the j th person in the i th school at the t th measurement We typically center the measurements at some point for convenience (often the middle of the time span) Centering strategy determines the interpretation of the coefficients of the growth model Note that the measurements do not have to be at exactly the same time for each person

Understanding a Polynomial Model Y ijk = β 0ij + β 1ij t ijk + β 2ij t ijk 2 + β 3ij t ijk 3 + ε ijk How do we interpret the coefficients? β 0ij is the intercept at the centering point β 1ij is the linear rate of growth at the centering point β 2ij is the acceleration (rate of change of linear growth) at the centering point β 3ij is the rate of change of the acceleration (often negative leading to flattening out of growth curves at the extremes)

Understanding a Polynomial Model Consider the quadratic growth model to understand acceleration with mean centering Thus you can see that the linear growth rate at time t is In other words, the linear growth rate increases with t and the only place where the linear growth rate is β 1ij is the middle

Understanding a Polynomial Model Thus β 1ij is the linear rate of growth at the centered value (here, the middle) If β 2ij > 0, the linear growth rate will be larger above the centered value and smaller below the centered value Centering at other values than the middle can make sense if that is where growth trajectory is of interest and if the model fits the data For example, centering at the end gives coefficients with interpretable rates at the end of the growth period

Understanding a Polynomial Model Consider the quadratic growth model to understand acceleration with mean centering Thus you can see that the acceleration at time t is In other words, the acceleration increases with t and the only place where the acceleration is β 2ij is the middle

Understanding a Polynomial Model Thus β 2ij is the acceleration of growth at the centered value (here the middle) If β 3ij > 0, the acceleration will be larger above the centered value and smaller below the centered value Centering at other values than the middle can make sense if that is where growth trajectory is of interest and if the model fits the data For example, centering at the end gives coefficients with interpretable rates at the end of the growth period

No Growth (Centered) β 0 = 5, β 1 = 0.00, β 2 = 0.00, β 3 = 0.00

Linear Growth (Centered) β 0 = 5, β 1 = 1, β 2 = 0.00, β 3 = 0.00

Quadratic Growth (Centered) β 0 = 5, β 1 = 1, β 2 = 0.05, β 3 = 0.00

Cubic Growth (Centered) β 0 = 5, β 1 = 1, β 2 = 0.05, β 3 = -0.01

Linear, Quadratic, and Cubic Growth (Centered) β 0 = 5, β 1 = 1, β 2 = 0.05, β 3 = -0.01,

Selecting Growth Models Several considerations are relevant in selecting a growth model First is how many repeated measures there are: The maximum degree is one less than the number of measures (linear needs 2, quadratic needs 3, etc.) However the estimates of growth parameters are much better if there are a few additional degrees of freedom But the most important consideration is whether the model fits the data! Unfortunately, this is not always completely unambiguous

Selecting Growth Models Individual growth trajectories are usually poorly estimated HLM models estimate average growth trajectories (via average parameters) and variation around that average: These are much more stable Estimates of individual growth curves can be greatly improved by using empirical Bayes methods to borrow strength from the average This may make sense if there all the individuals in the groups are sampled from a common population It can be problematic if some individuals are dramatically different

Selecting Analysis Models One issue is selecting the growth model to characterize growth A different, but related, issue is selecting how treatment should impact growth Should it impact linear growth term? Should it impact the acceleration? Which impact is primary? How does looking at multiple impacts weaken the design? What if impacts are in opposite directions?

Longitudinal Experiments Assigning Treatment to Schools In the language of experimental design, adding repeated measures adds another factor to the design: A measures factor The measures factor is crossed with individuals, treatments, and clusters Schools are nested within the treatment factor and individuals are nested within school by treatments Repeated measures analysis of variance can be used to analyze these designs, but we will not pursue that point of view Instead we will use the HLM notation

Longitudinal Experiments Assigning Treatment To Schools Level 1 (measures) Y ijk = β 0ij + β 1ij t ijk + β 2ij t ijk 2 + ε ijk Level 2 (individuals) β 0ij = γ 00j + η 0ij η ~ N(0, Σ I ) β 1ij = γ 10j + η 1ij β 2ij = γ 20j + η 2ij Level 3 (schools) γ 00j = π 00 + π 01 T i + ξ 0j ξ ~ N(0, Σ S ) γ 01j = π 10 + π 11 T i + ξ 1j γ 20j = π 20 + π 21 T i + ξ 2j

Longitudinal Experiments Assigning Treatment To Schools This model has three trend coefficients in each growth trajectory Note that there are 3 random effects at the second and third level This means that 6 variances and covariances must be estimated at each level This may require more information to do accurately than is available at the school level It is often prudent to fix some of these effects because they cannot all be estimated accurately

Longitudinal Experiments Assigning Treatment Within Schools In the language of experimental design, adding repeated measures adds another factor to the design: A measures factor The measures factor is crossed with individuals, treatments, and clusters The treatment factor is crossed with schools and individuals are nested within school by treatments Repeated measures analysis of variance can be used to analyze these designs, but we will not pursue that point of view

Longitudinal Experiments Assigning Treatment Within Schools Level 1 (measures level) Y ijk = β 0ij + β 1ij t + β 2ij t 2 + ε ijk ε ~ N(0, σ W 2 ) Level 2 (individual level) β 0ij = γ 00j + γ 01j T j + η 0ij η ~ N(0, Σ C ) β 1ij = γ 10j + γ 11j T j + η 1ij β 2ij = γ 20j + γ 21j T j + η 2ij Level 3 (school level) γ 00j = π 00 + ξ 00j ξ a0 ~ N(0, Σ S ) γ 01j = π 10 + ξ 01j ξ a1 ~ N(0, Σ T x S ) γ 10j = π 00 + ξ 10j γ 11j = π 00 + ξ 11j γ 20j = π 00 + ξ 20j γ 21j = π 00 + ξ 21j

Longitudinal Experiments Assigning Treatment Within Schools This model has three trend coefficients in each growth trajectory Note that there are 6 random effects at the third level This means that 15 variances and covariances must be estimated at the third level This requires a great deal of information to do accurately It is often prudent to fix some of these effects because they cannot all be estimated accurately However there is some art in this, and sensitivity analysis is a good precaution

Longitudinal Experiments Covariates can be added at any level of the design But remember that covariates must be variables that cannot have been impacted by treatment assignment Thus time varying covariates (at level 1) are particularly suspect since they may be measured after treatment assignment

Power Analysis Power computations for longitudinal experiments are doable, but depend on parameters that may not be well known For example reliability of trend coefficients When parameters such as these are known, the computations are straightforward, but there is relatively little information about them that can be used for planning To make matters worse, the values of some parameters (such as reliability) depend on the number of measures Thus it is often necessary to rely on values of variance components

Power Analysis Still some generalizations are possible Power increases with the number of measures Power increases with the length of time over which measures are made (except for β 0ij ) Power increases with the precision of each individual measure These factors impact different trend coefficients differently Clustering increases the complexity of computations

Power Analysis Pilot data (or data from related studies, perhaps non-experimental ones) is more important in planning longitudinal experiments Longitudinal experiments to look at growth trajectories are attractive, but this is an area at the frontier of practical experience Research is ongoing to produce better methods for power analysis of longitudinal experiments that will be practically useful

Good Luck!