Inference: Neyman’s Repeated Sampling STA 320 Design and Analysis of Causal Studies Dr. Kari Lock Morgan and Dr. Fan Li Department of Statistical Science.

Slides:



Advertisements
Similar presentations
Inference: Fishers Exact p-values STA 320 Design and Analysis of Causal Studies Dr. Kari Lock Morgan and Dr. Fan Li Department of Statistical Science Duke.
Advertisements

Comparing One Sample to its Population
Simulating with StatKey Kari Lock Morgan Department of Statistical Science Duke University Joint Mathematical Meetings, San Diego 1/11/13.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 21, Slide 1 Chapter 21 Comparing Two Proportions.
Objectives (BPS chapter 18) Inference about a Population Mean  Conditions for inference  The t distribution  The one-sample t confidence interval 
Review of the Basic Logic of NHST Significance tests are used to accept or reject the null hypothesis. This is done by studying the sampling distribution.
Inference for a population mean BPS chapter 18 © 2006 W. H. Freeman and Company.
Hypothesis Testing: Hypotheses
Inferential Statistics & Hypothesis Testing
Review: What influences confidence intervals?
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Sampling Distributions
Topic 2: Statistical Concepts and Market Returns
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 9: Hypothesis Tests for Means: One Sample.
BHS Methods in Behavioral Sciences I
Lecture Inference for a population mean when the stdev is unknown; one more example 12.3 Testing a population variance 12.4 Testing a population.
Horng-Chyi HorngStatistics II41 Inference on the Mean of a Population - Variance Known H 0 :  =  0 H 0 :  =  0 H 1 :    0, where  0 is a specified.
Chapter 11: Inference for Distributions
1 Inference About a Population Variance Sometimes we are interested in making inference about the variability of processes. Examples: –Investors use variance.
5-3 Inference on the Means of Two Populations, Variances Unknown
Standard error of estimate & Confidence interval.
Introducing Inference with Simulation Methods; Implementation at Duke University Kari Lock Morgan Department of Statistical Science, Duke University
Weighting STA 320 Design and Analysis of Causal Studies Dr. Kari Lock Morgan and Dr. Fan Li Department of Statistical Science Duke University.
Statistics: Unlocking the Power of Data Lock 5 Hypothesis Testing: Hypotheses STAT 101 Dr. Kari Lock Morgan SECTION 4.1 Statistical test Null and alternative.
AP Statistics Section 13.1 A. Which of two popular drugs, Lipitor or Pravachol, helps lower bad cholesterol more? 4000 people with heart disease were.
Ch 10 Comparing Two Proportions Target Goal: I can determine the significance of a two sample proportion. 10.1b h.w: pg 623: 15, 17, 21, 23.
Ch 11 – Inference for Distributions YMS Inference for the Mean of a Population.
1 Level of Significance α is a predetermined value by convention usually 0.05 α = 0.05 corresponds to the 95% confidence level We are accepting the risk.
More About Significance Tests
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Statistical Inferences Based on Two Samples Chapter 9.
BPS - 3rd Ed. Chapter 211 Inference for Regression.
Hypothesis Testing: p-value
Dan Piett STAT West Virginia University
Comparing Two Proportions
1 Introduction to Estimation Chapter Concepts of Estimation The objective of estimation is to determine the value of a population parameter on the.
Estimates and Sample Sizes Lecture – 7.4
Chapter 10 Comparing Two Means Target Goal: I can use two-sample t procedures to compare two means. 10.2a h.w: pg. 626: 29 – 32, pg. 652: 35, 37, 57.
Introducing Inference with Simulation Methods; Implementation at Duke University Kari Lock Morgan Department of Statistical Science, Duke University
Education Research 250:205 Writing Chapter 3. Objectives Subjects Instrumentation Procedures Experimental Design Statistical Analysis  Displaying data.
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
AP Statistics Section 13.1 A. Which of two popular drugs, Lipitor or Pravachol, helps lower bad cholesterol more? 4000 people with heart disease were.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Analysis of Variance (ANOVA) Can compare the effects of different treatments Can make population level inferences based on sample population.
Nonparametric Statistics
Introducing Inference with Bootstrapping and Randomization Kari Lock Morgan Department of Statistical Science, Duke University with.
Copyright © 2010 Pearson Education, Inc. Chapter 22 Comparing Two Proportions.
Statistics: Unlocking the Power of Data Lock 5 Bootstrap Intervals Dr. Kari Lock Morgan PSU /12/14.
ANOVA Assumptions 1.Normality (sampling distribution of the mean) 2.Homogeneity of Variance 3.Independence of Observations - reason for random assignment.
Statistics: Unlocking the Power of Data Lock 5 Exam 2 Review STAT 101 Dr. Kari Lock Morgan 11/13/12 Review of Chapters 5-9.
Ch11: Comparing 2 Samples 11.1: INTRO: This chapter deals with analyzing continuous measurements. Later, some experimental design ideas will be introduced.
1 URBDP 591 A Lecture 12: Statistical Inference Objectives Sampling Distribution Principles of Hypothesis Testing Statistical Significance.
Review Normal Distributions –Draw a picture. –Convert to standard normal (if necessary) –Use the binomial tables to look up the value. –In the case of.
Review of Statistics.  Estimation of the Population Mean  Hypothesis Testing  Confidence Intervals  Comparing Means from Different Populations  Scatterplots.
Chapter 22 Comparing Two Proportions.  Comparisons between two percentages are much more common than questions about isolated percentages.  We often.
Hypothesis Testing. Suppose we believe the average systolic blood pressure of healthy adults is normally distributed with mean μ = 120 and variance σ.
Inference about proportions Example: One Proportion Population of students Sample of 175 students CI: What proportion (percentage) of students abstain.
Statistics: Unlocking the Power of Data Lock 5 Hypothesis Testing: p-value STAT 250 Dr. Kari Lock Morgan SECTION 4.2 p-value.
BPS - 5th Ed. Chapter 231 Inference for Regression.
Chapter 9 Sampling Distributions 9.1 Sampling Distributions.
Statistics: Unlocking the Power of Data Lock 5 Section 4.1 Introducing Hypothesis Tests.
Chapter 11 Inference for Distributions AP Statistics 11.2 – Inference for comparing TWO Means.
Chapter 9 Roadmap Where are we going?.
Parameter Estimation.
Chapter 4. Inference about Process Quality
Introducing Hypothesis Tests
Comparing Two Proportions
Comparing Two Proportions
Basic Practice of Statistics - 3rd Edition Inference for Regression
Estimates and Sample Sizes Lecture – 7.4
Presentation transcript:

Inference: Neyman’s Repeated Sampling STA 320 Design and Analysis of Causal Studies Dr. Kari Lock Morgan and Dr. Fan Li Department of Statistical Science Duke University

Office Hours My Monday office hours will be 12-1pm for the next 3 weeks (2/10, 2/17, 2/24), not 3-4pm Wednesdays: 3-4pm Fridays: 1-3pm

R R code corresponding to all the problems from last class is available R code There are many different ways to code each problem – this is just an example For more information on any of it (for loops, subsetting data, handling NAs, etc.) see this R guide I wrote for intro statR guide You will have to do your own coding for homework and your project (you can talk, but do not share code)

HW 2 Because the due date for HW 2 has gotten pushed back a week (now due Monday, 2/10), the next hw was dropped and instead I’ve added some problems to HW corresponding to today’s class If you already downloaded it, make sure to look at the updated version

Causal Inference

Sleep or Caffeine? Is sleep or caffeine better for memory? 24 adults were given a list of words to memorize, then randomly divided into two groups During a break one group took a nap for an hour and a half, while the other group stayed awake and then took a caffeine pill after an hour Y: number of words recalled Mednick S., Cai D., Kanady J., and Drummond S., “Comparing the benefits of caffeine, naps and placebo on verbal, motor and perceptual memory”, Behavioural Brain Research, 2008; 193:

Sleep or Caffeine

Jerzy Neyman 1894 – 1981

Fisher and Neyman At the same time Fisher was developing his framework for inference, Neyman was developing his own framework… Fisher: more focused on testing o is there a difference? o p-values Neyman: more focused on estimation o average treatment effect o unbiased estimators o confidence intervals

Sleep or Caffeine Fisher: Is there any difference between napping or staying awake and consuming caffeine, regarding number of words recalled? Neyman: On average, how many more words are recalled if a person naps rather than stays awake and consumes caffeine?

Neyman’s Plan for Inference 1.Define the estimand 2.Look for an unbiased estimator of the estimand 3.Calculate the true sampling variance of the estimator 4.Look for an unbiased estimator of the true sampling variance of the estimator 5.Assume approximate normality to obtain p-value and confidence interval 11 Slide adapted from Cassandra Pattanayak, Harvard

Finite Sample vs Super Population Finite sample inference: o Only concerned with units in the sample o Only source of randomness is random assignment to treatment groups o (Fisher exact p-values) Super population inference: o Extend inferences to greater population o Two sources of randomness: random sampling, random assignment o “repeated sampling” We’ll first explore finite sample inference…

Estimand Neyman was primarily interested in estimating the average treatment effect In the finite sample setting, this is defined as

Estimator A natural estimator is the difference in observed sample means:

Sleep vs Caffeine Estimand: the average word recall for all 24 people if they had napped – average word recall for all 24 people if they had caffeine Estimator: (Sleep – Caffeine)

Unbiased An estimator is unbiased is the average of the estimator computed over all assignment vectors (W) will equal the estimand The estimator is unbiased if

Unbiased For completely randomized experiments, is an unbiased estimator for

Neyman’s Inference (Finite Sample) 1.Define the estimand: 2. unbiased estimator of the estimand: 1.Calculate the true sampling variance of the estimator 18 Slide adapted from Cassandra Pattanayak, Harvard

True Variance over W For the derivation of this, see Chapter 6. Sample variance of potential outcomes under treatment and control, for all units.

Extra Term Always positive Equal to zero if the treatment effect is constant for all i Related to the correlation between Y(0) and Y(1), (perfectly correlated if constant treatment effect)

Neyman’s Inference (Finite Sample) 1.Define the estimand: 2. unbiased estimator of the estimand: 3. true sampling variance of the estimator 4.Look for an unbiased estimator of the true sampling variance of the estimator 21 Slide adapted from Cassandra Pattanayak, Harvard (IMPOSSIBLE!)

Estimator of Variance (of estimator) Sample variances of observed outcomes under treatment and control (look familiar???)

Estimator of Variance This is the standard variance estimate used in the familiar t-test For finite samples, this is may be an overestimate of the true variance Resulting inferences may be too conservative (confidence intervals will be too wide, p-values too large)

Sleep vs Caffeine

Neyman’s Inference (Finite Sample) 1.Define the estimand: 2. unbiased estimator of the estimand: 3. true sampling variance of the estimator 4. unbiased estimator of the true sampling variance of the estimator 5.Assume approximate normality to obtain p- value and confidence interval 25 (IMPOSSIBLE!) Overestimate: Slide adapted from Cassandra Pattanayak, Harvard

Central Limit Theorem Neyman’s inference relies on the central limit theorem: sample sizes must be large enough for the distribution of the estimator to be approximately normal Depends on sample size AND distribution of the outcome (need larger N if highly skewed, outliers, or rare binary events)

Confidence Intervals z* (or t*) is the value leaving the desired percentage in between –z* and z* in the standard normal distribution (Confidence intervals due to Neyman!)

Confidence Intervals For finite sample inference: Intervals may be too wide Inference may be too conservative A 95% interval will contain the estimand at least 95% of the time

Sleep vs Caffeine > qt(.975, df=11) [1] % CI: (-0.08, 6.08)

Confidence Intervals - Fisher You can also get confidence intervals from inverting the Fisher randomization test Rather than assuming no treatment effect, assume a constant treatment effect, c, and do a randomization test The 95% confidence interval is all values of c that would not be rejected at the 5% significance level

Hypothesis Testing Fisher: sharp null hypothesis of no treatment effect for any unit Neyman: null hypothesis of no treatment effect on average

Hypothesis Testing Fisher: compare any test statistic to empirical randomization distribution Neyman: compare t-statistic to normal or t distribution (relies on large n) (Neyman’s approach is the familiar t-test)

Sleep vs Caffeine > pt(2.14, df=11, lower.tail=FALSE) [1]

Sleep vs Caffeine Exact p-value =

Super Population Suppose we also want to consider random sampling from the population (in addition to random assignment) How do things change?

Neyman Inference (Super Population) 1.Define the estimand: 2. unbiased estimator of the estimand: 3. true sampling variance of the estimator 4. unbiased estimator of the true sampling variance of the estimator 5.Assume approximate normality to obtain p- value and confidence interval 36 Slide adapted from Cassandra Pattanayak, Harvard

Super Population Neyman’s results (and therefore all the familiar t-based inference you are used to) are considering both random sampling from the population and random assignment

Fisher vs Neyman FisherNeyman Goal: testingGoal: estimation Considers only random assignment Considers random assignment and random sampling H 0 : no treatment effectH 0 : average treatment effect = 0 Works for any test statistic Difference in means Exact distributionApproximate, relies on large n Works for any known assignment mechanism Only derived for common designs

To Do Read Ch 6 HW 2 due Monday 2/10