Power Analysis: Sample Size, Effect Size, Significance, and Power 16 Sep 2011 Dr. Sean Ho CPSY501 cpsy501.seanho.com Please download: SpiderRM.sav.

Slides:



Advertisements
Similar presentations
Inference for Regression
Advertisements

Lecture 3: Null Hypothesis Significance Testing Continued Laura McAvinue School of Psychology Trinity College Dublin.
CPSY 501: Class 2 Outline Please log-in and download the lec-2 data set from the class web-site. Statistical significance Effect size Combining effect.
RIMI Workshop: Power Analysis Ronald D. Yockey
ANOVA: PART II. Last week  Introduced to a new test:  One-Way ANOVA  ANOVA’s are used to minimize family-wise error:  If the ANOVA is statistically.
Using Statistics in Research Psych 231: Research Methods in Psychology.
Using Statistics in Research Psych 231: Research Methods in Psychology.
Lecture 11 Psyc 300A. Null Hypothesis Testing Null hypothesis: the statistical hypothesis that there is no relationship between the variables you are.
1 Practicals, Methodology & Statistics II Laura McAvinue School of Psychology Trinity College Dublin.
Lecture 9: One Way ANOVA Between Subjects
PSY 1950 Confidence and Power December, Requisite Quote “The picturing of data allows us to be sensitive not only to the multiple hypotheses that.
Simple Correlation Scatterplots & r Interpreting r Outcomes vs. RH:
UNDERSTANDING RESEARCH RESULTS: STATISTICAL INFERENCE © 2012 The McGraw-Hill Companies, Inc.
Using Statistics in Research Psych 231: Research Methods in Psychology.
Using Statistics in Research Psych 231: Research Methods in Psychology.
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Basic Relationships Purpose of multiple regression Different types of multiple regression.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Split Sample Validation General criteria for split sample validation Sample problems.
Inferential Statistics
Practical statistics for Neuroscience miniprojects Steven Kiddle Slides & data :
Statistical Analyses & Threats to Validity
Selecting the Correct Statistical Test
LEARNING PROGRAMME Hypothesis testing Intermediate Training in Quantitative Analysis Bangkok November 2007.
Choosing and using statistics to test ecological hypotheses
User Study Evaluation Human-Computer Interaction.
Statistics (cont.) Psych 231: Research Methods in Psychology.
Linear Regression 1 Oct 2010 CPSY501 Dr. Sean Ho Trinity Western University Please download from “Example Datasets”: Record2.sav ExamAnxiety.sav.
Statistics (cont.) Psych 231: Research Methods in Psychology.
Hypothesis testing Intermediate Food Security Analysis Training Rome, July 2010.
Intro: “BASIC” STATS CPSY 501 Advanced stats requires successful completion of a first course in psych stats (a grade of C+ or above) as a prerequisite.
SPSS Basics and Applications Workshop: Introduction to Statistics Using SPSS.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Sampling, sample size estimation, and randomisation
Categorical Data Analysis: Logistic Regression and Log-Linear Regression 26 Nov 2010 CPSY501 Dr. Sean Ho Trinity Western University For discussion: Myers.
SW388R6 Data Analysis and Computers I Slide 1 Multiple Regression Key Points about Multiple Regression Sample Homework Problem Solving the Problem with.
12: Basic Data Analysis for Quantitative Research.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Ch10: T-tests 6 Mar 2012 Dr. Sean Ho busi275.seanho.com Please download: 08-TTests.xls 08-TTests.xls HW5 this week Projects.
Mixed-Design ANOVA 5 Nov 2010 CPSY501 Dr. Sean Ho Trinity Western University Please download: treatment5.sav.
Experimental Research Methods in Language Learning Chapter 10 Inferential Statistics.
Lab 9: Two Group Comparisons. Today’s Activities - Evaluating and interpreting differences across groups – Effect sizes Gender differences examples Class.
The “Big 4” – Significance, Effect Size, Sample Size, and Power 18 Sep 2009 Dr. Sean Ho CPSY501 cpsy501.seanho.com.
CPSY 501: Advanced Statistics 11 Sep 2009 Instructor: Dr. Sean Ho TA: Marci Wagner cpsy501.seanho.com No food/drink in the computer lab, please! Please.
Chapter 6: Analyzing and Interpreting Quantitative Data
Tuesday, April 8 n Inferential statistics – Part 2 n Hypothesis testing n Statistical significance n continued….
Inferential Statistics. Explore relationships between variables Test hypotheses –Research hypothesis: a statement of the relationship between variables.
: Pairwise t-Test and t-Test on Proportions 25 Oct 2011 BUSI275 Dr. Sean Ho HW6 due Thu Please download: Mileage.xls Mileage.xls.
Handout Eight: Two-Way Between- Subjects Design with Interaction- Assumptions, & Analyses EPSE 592 Experimental Designs and Analysis in Educational Research.
Research Methods and Data Analysis in Psychology Spring 2015 Kyle Stephenson.
1 Week 3 Association and correlation handout & additional course notes available at Trevor Thompson.
Difference Between Means Test (“t” statistic) Analysis of Variance (F statistic)
Statistics (cont.) Psych 231: Research Methods in Psychology.
Chapter 13 Understanding research results: statistical inference.
Mixed-Design ANOVA 13 Nov 2009 CPSY501 Dr. Sean Ho Trinity Western University Please download: treatment5.sav.
Statistics (cont.) Psych 231: Research Methods in Psychology.
Beginners statistics Assoc Prof Terry Haines. 5 simple steps 1.Understand the type of measurement you are dealing with 2.Understand the type of question.
Analyzing Data. Learning Objectives You will learn to: – Import from excel – Add, move, recode, label, and compute variables – Perform descriptive analyses.
Inferential Statistics Psych 231: Research Methods in Psychology.
Choosing and using your statistic. Steps of hypothesis testing 1. Establish the null hypothesis, H 0. 2.Establish the alternate hypothesis: H 1. 3.Decide.
Appendix I A Refresher on some Statistical Terms and Tests.
Inferential Statistics Psych 231: Research Methods in Psychology.
Other tests of significance. Independent variables: continuous Dependent variable: continuous Correlation: Relationship between variables Regression:
CHAPTER 15: THE NUTS AND BOLTS OF USING STATISTICS.
UNDERSTANDING RESEARCH RESULTS: STATISTICAL INFERENCE
Inferential Statistics
Multiple Regression – Split Sample Validation
Psych 231: Research Methods in Psychology
Psych 231: Research Methods in Psychology
Psych 231: Research Methods in Psychology
Psych 231: Research Methods in Psychology
Presentation transcript:

Power Analysis: Sample Size, Effect Size, Significance, and Power 16 Sep 2011 Dr. Sean Ho CPSY501 cpsy501.seanho.com Please download: SpiderRM.sav

16 Sep 2011CPSY501: power analysis2 Outline for today Discussion of research article (Missirlian et al) Power analysis (the “Big 4”): Statistical significance (p-value, α )  What does the p-value mean, really? Power (1- β ) Effect size (Cohen's rules of thumb) Sample size (n)  Calculating min required sample size Overview of linear models for analysis SPSS tips for data prep and analysis

16 Sep 2011CPSY501: power analysis3 Practice reading article Last time you were asked to read this journal article, focusing on the statistical methods: Missirlian, et al., “Emotional Arousal, Client Perceptual Processing, and the Working Alliance in Experiential Psychotherapy for Depression”, Journal of Consulting and Clinical Psychology, Vol. 73, No. 5, pp. 861–871, How much were you able to understand?

16 Sep 2011CPSY501: power analysis4 Discussion: What research questions do the authors state that they are addressing? What analytical strategy was used, and how appropriate is it for addressing their questions? What were their main conclusions, and are these conclusions warranted from the actual results /statistics /analyses that were reported? What, if any, changes/additions need to be made to the methods to give a more complete picture of the phenomenon of interest (e.g., sampling, description of analysis process, effect sizes, dealing with multiple comparisons, etc.)?

16 Sep 2011CPSY501: power analysis5 Outline for today Discussion of research article (Missirlian et al) Power analysis (the “Big 4”): Statistical significance (p-value, α )  What does the p-value mean, really? Power (1- β ) Effect size (Cohen's rules of thumb) Sample size (n)  Calculating min required sample size Overview of linear models for analysis SPSS tips for data prep and analysis

16 Sep 2011CPSY501: power analysis6 Central Themes of Statistics Is there a real effect/relationship amongst the given variables? How big is that effect? We evaluate these by looking at Statistical significance (p-value) and Effect size (r 2, R 2, η, d, etc.) Along with sample size (n) and statistical power (1-β), these form the “Big 4” of any test

16 Sep 2011CPSY501: power analysis7 The “Big 4” of every test Any statistical test has these 4 facets Set 3 as desired → derive required level of 4 th Significance Sample sizeEffect size Power 0.05 as desired, depends on test usually 0.80 derived! (GPower)

16 Sep 2011CPSY501: power analysis8 Significance ( α ) vs. power (1- β ) α is our tolerance of Type-I error: incorrectly rejecting the null hypothesis (H 0 ) β is our tolerance of Type-II error: failing to reject H 0 when we should have rejected H 0. Set H 0 so that Type-I is the worse kind of error Power is 1- β e.g., parachute inspections (what should H 0 be?) Actual truth H 0 trueH 0 false Reject H 0 Type-I error (α) Correct Fail to reject H 0 Correct Type-II error (β) O ur d ec is io n

16 Sep 2011CPSY501: power analysis9 Significance: example RQ: are depression levels higher in males than in females? IV: gender; DV: depression (e.g., DES) What is H 0 ? (recall: H 0 means no effect!) Analysis: independent-groups t-test What does Type-I error mean? What does Type-II error mean?

16 Sep 2011CPSY501: power analysis10 Impact of changing α α =0.05 α =0.01 (without changing data) decreasing Type-I means increasing Type-II!

16 Sep 2011CPSY501: power analysis11 What is the p-value? Given data and assuming a model, the p-value is the computed probability of Type-I error Assuming model is true and H 0 is true, p is the probability of getting data that is at least as extreme as what we observed Note: this does not tell us the probability that H 0 is true! p(data|H 0 ) vs. p(H 0 |data) Set the significance level ( α =0.05) according to our tolerance for Type-I error: if p < α, we are confident enough to say the effect is real

16 Sep 2011CPSY501: power analysis12 Myths about significance (why are these all myths?) Myth 1: “If a result is not significant, it proves there is no effect.” Myth 2: “The obtained significance level indicates the reliability of the research finding.” Myth 3: “The significance level tells you how big or important an effect is.” Myth 4: “If an effect is statistically significant, it must be clinically significant.”

16 Sep 2011CPSY501: power analysis13 Problems with relying on p-val When sample size is low, p is usually too big → if effect size is big, try bigger sample When sample size is very big, p can easily be very small even for tiny effects e.g., avg IQ of men is 0.8pts higher than IQ of women, in a sample of 10,000: statistically significant, but is it clinically significant? When many tests are run, one of them is bound to turn up significant by random chance → multiple comparisons: inflated Type-I

16 Sep 2011CPSY501: power analysis14 Outline for today Discussion of research article (Missirlian et al) Power analysis (the “Big 4”): Statistical significance (p-value, α )  What does the p-value mean, really? Power (1- β ) Effect size (Cohen's rules of thumb) Sample size (n)  Calculating min required sample size Overview of linear models for analysis SPSS tips for data prep and analysis

16 Sep 2011CPSY501: power analysis15 Effect size Historically, researchers only looked at significance, but what about the effect size? A small study might yield non-significance but a strong effect size Could be spurious, but could also be real Motivates meta-analysis – repeat the experiment, combine results: with more data, it might be significant! Current research standards require reporting both p-value as well as effect size

16 Sep 2011CPSY501: power analysis16 Measures of effect size For t-test and any between-groups comparison: Difference of means: d = ( μ 1 – μ 2 )/ σ For ANOVA: η 2 (eta-squared): Overall effect of IV on DV For bivariate correlation (Pearson, Spearman): r or r 2 : r 2 is fraction of variability in one var explained by the other var For regression: R 2 and ΔR 2 (“R 2 -change”) Fraction of variability in DV explained by overall model (R 2 ) or by each predictor (ΔR 2 )

16 Sep 2011CPSY501: power analysis17 Interpreting effect size What constitutes a “big” effect size? Consult literature for the phenomenon of study Cohen '92, “A Power Primer”: rules of thumb Somewhat arbitrary, though! For correlation-type r measures: 0.10 → small effect (1% of var. explained) 0.30 → medium effect (9% of variability) 0.50 → large effect (25%)

16 Sep 2011CPSY501: power analysis18 Example: dependent t-test Dataset: SpiderRM.sav 12 individuals, first shown picture of spider, then shown real spider → measured anxiety Compare Means → Paired-Samples T Test SPSS results: t(11)=-2.473, p<0.05 Calculate effect size: see text, p.332 ( § 9.4.6) r ~ (big? Small?) Report sample size (df), test statistic (t), p-value, and effect size (r)

16 Sep 2011CPSY501: power analysis19 Finding needed sample size Experimental design: what is the minimum sample size needed to attain the desired level of significance, power, and effect size (assuming there is a real relationship)? Choose level of significance: usually α = 0.05 Choose power: usually 1- β = 0.80 Choose desired effect size From literature or Cohen's rules of thumb → use GPower or similar to calculate the required sample size (do this for your project!)

16 Sep 2011CPSY501: power analysis20 Power vs. sample size Choose α =0.05, effect size d=0.50, and vary β :

16 Sep 2011CPSY501: power analysis21 Outline for today Discussion of research article (Missirlian et al) Power analysis (the “Big 4”): Statistical significance (p-value, α )  What does the p-value mean, really? Power (1- β ) Effect size (Cohen's rules of thumb) Sample size (n)  Calculating min required sample size Overview of linear models for analysis SPSS tips for data prep and analysis

16 Sep 2011CPSY501: power analysis22 Overview of Linear Methods For scale-level DV: dichotomous IV → indep t-test categorical IV → one-way ANOVA multiple categorical IVs → factorial ANOVA scale IV → correlation or regression multiple scale IVs → multiple regression For categorical DV: categorical IV → χ 2 or log-linear scale IV → logistic regression Also: repeated measures and mixed-effects

16 Sep 2011CPSY501: power analysis23 SPSS tips! Plan out the characteristics of your variables before you start data-entry You can reorder variables: cluster related ones Create a var holding unique ID# for each case Variable names can't have spaces: try “_” Add descriptive labels to your variables Code missing data using values like '999' Then tell SPSS about it in Variable View Clean up your output file (*.spv) before turn-in Add text boxes / headers; delete junk