If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.

Slides:



Advertisements
Similar presentations
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Advertisements

Chapter 10 Introduction to Inference
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Quantitative Data Analysis
A Spreadsheet for Analysis of Straightforward Controlled Trials
Research Skills Workshop Designing a Project
Client Assessment and Other New Uses of Reliability Will G Hopkins Physiology and Physical Education University of Otago, Dunedin NZ Reliability: the Essentials.
Statistical vs Clinical or Practical Significance
Planning, Performing, and Publishing Research with Confidence Limits
Statistical vs Clinical Significance
Validity and Reliability
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Equivalence Testing Dig it!.
Designing an impact evaluation: Randomization, statistical power, and some more fun…
Power and sample size.
Type I & Type II errors Brian Yuen 18 June 2013.
CHAPTER 14: Confidence Intervals: The Basics
Statistical POWERAnalysis Statistical POWER Analysis : A Conceptual Tutorial Experimental Psychology Course Dr. Morre Goldsmith March, 2005 Copyright ©
Hypothesis testing Another judgment method of sampling data.
Statistical Analysis and Data Interpretation What is significant for the athlete, the statistician and team doctor? important Will Hopkins
Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Statistics.  Statistically significant– When the P-value falls below the alpha level, we say that the tests is “statistically significant” at the alpha.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
Issues About Statistical Inference Dr R.M. Pandey Additional Professor Department of Biostatistics All-India Institute of Medical Sciences New Delhi.
1 1 Slide STATISTICS FOR BUSINESS AND ECONOMICS Seventh Edition AndersonSweeneyWilliams Slides Prepared by John Loucks © 1999 ITP/South-Western College.
MAGNITUDE-BASED INFERENCES
Statistical Issues in Research Planning and Evaluation
Review: What influences confidence intervals?
T-tests Computing a t-test  the t statistic  the t distribution Measures of Effect Size  Confidence Intervals  Cohen’s d.
Statistical Tests How to tell if something (or somethings) is different from something else.
Example 10.1 Experimenting with a New Pizza Style at the Pepperoni Pizza Restaurant Concepts in Hypothesis Testing.
Are the results valid? Was the validity of the included studies appraised?
Overview Definition Hypothesis
Statistical Techniques I
Hypothesis Testing.
Confidence Intervals and Hypothesis Testing
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
1 Statistical Inference. 2 The larger the sample size (n) the more confident you can be that your sample mean is a good representation of the population.
Chapter 8 Introduction to Hypothesis Testing
Comparing two sample means Dr David Field. Comparing two samples Researchers often begin with a hypothesis that two sample means will be different from.
Instructor Resource Chapter 5 Copyright © Scott B. Patten, Permission granted for classroom use with Epidemiology for Canadian Students: Principles,
 If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
PowerPoint presentation to accompany Research Design Explained 6th edition ; ©2007 Mark Mitchell & Janina Jolley Chapter 10 The Simple Experiment.
Statistical Power The power of a test is the probability of detecting a difference or relationship if such a difference or relationship really exists.
Chapter 8 Delving Into The Use of Inference 8.1 Estimating with Confidence 8.2 Use and Abuse of Tests.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
KNR 445 Statistics t-tests Slide 1 Introduction to Hypothesis Testing The z-test.
Statistics for Decision Making Basic Inference QM Fall 2003 Instructor: John Seydel, Ph.D.
Research planning. Planning v. evaluating research To a large extent, the same thing Plan a study so that it is capable of yielding data that could possibly.
P Values - part 2 Samples & Populations Robin Beaumont 2011 With much help from Professor Chris Wilds material University of Auckland.
Chapter 13 Understanding research results: statistical inference.
The inference and accuracy We learned how to estimate the probability that the percentage of some subjects in the sample would be in a given interval by.
Making Inferences About Effects Seminar presented at Leeds Beckett and Split universities, March 2016 This slideshow consists of part of the lecture on.
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
Chapter 9 Introduction to the t Statistic
Critical Appraisal Course for Emergency Medicine Trainees Module 2 Statistics.
Inference: Conclusion with Confidence
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Introduction to inference Use and abuse of tests; power and decision
Hypothesis Testing: Preliminaries
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Review: What influences confidence intervals?
I figured there must be a better way. There is: confidence intervals!
Hypothesis Testing.
Quantitative Data Analysis
Sample Sizes for IE Power Calculations.
Presentation transcript:

If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it directly in PowerPoint. When you open the file, use the full-screen view to see the information on each slide build sequentially. For full-screen view, click on this icon at the lower left of your screen. To go forwards, left-click or hit the space bar, PdDn or key. To go backwards, hit the PgUp or key. To exit from full-screen view, hit the Esc (escape) key.

Will G Hopkins Sport and Recreation AUT University Auckland NZ Clinically or Practically Decisive Sample Sizes General Principles Sample vs population Ethics Effects of effect magnitude, design, validity, reliability Approaches to Sample-Size Estimation What others have used Statistical significance Precision of estimation Clinical decisiveness

General Principles We study an effect in a sample, but we want to know about the effect in the population. The larger the sample, the closer we get to the population. Too large is unethical, because it's wasteful. Too small is unethical, because the effect won't be clear. And you are less likely to get your study published. But meta-analysis of several such studies leads to a clear outcome, so small-scale studies should be published. The bigger the effect, the smaller the sample you need to get a clear effect. So start with a smallish sample, then add more if necessary. But this approach may overestimate the effect.

More General Principles Sample size depends on the design. Cross-sectional studies (case-control, correlational) usually need hundreds of subjects. Controlled-trial interventions usually need scores of subjects. Crossover interventions usually need 10 or so subjects. Sample size depends on the validity (for cross-sectional studies) and reliability (for trials). Different approaches to estimation of sample size give different estimates of sample size. Traditional approach #1: what others have used. Traditional approach #2: statistical significance. Newer approach: acceptable precision of estimation. Newest approach: clinical decisiveness.

Traditional Approach #1: Use What Others Have Used No-one will believe your study in isolation, no matter what the sample size. A meta-analyst will combine your study with others, so... You might as well use the sample size that others have used, because… If the journal editors accepted their studies, they should accept yours. But your measurements need to be comparable to what others have used. Example: if your measure is less reliable, your outcome will be less clear unless you use more subjects.

Traditional Approach #2: Statistical Significance You need enough subjects to "detect" (get statistical significance for) the smallest important effect most of the time. You set a Type I error rate = chance of detecting null effect (5%) and a Type II error rate = chance of missing smallest effect (20%) or power = chance of detecting smallest effect = % = 80%. Problem: statistical non/significance is easy to misinterpret. Problem: this approach leads to large sample sizes. Example: ~800 subjects for case-control study to detect a standardized (Cohen) effect size of 0.2 or a correlation of 0.1. Samples are even larger if you keep the overall p<0.05 for multiple effects (the problem of inflation of Type 1 error). Smaller samples give clinically or practically decisive outcomes in our discipline.

The Type I error rate (5%) defines a critical value of the statistic. If observed value > critical value, the effect is significant. SIGNIFICANT NON-SIG Statistical Significance: How It Works area = 2.5% area = 20% When true value = smallest important value, the Type II error rate (20%) = chance of observing non-significant values. Solve for the sample size (via the critical value). probability value of effect statistic 0 positivenegative distribution of observed values, if true value = smallest important value distribution of observed values, if true value = 0 critical value smallest important value

Newer Approach: Acceptable Precision of Estimation Many researchers now report precision of estimation using confidence limits. Confidence limits define a range within which the true value of the effect is likely to be. Therefore they should justify sample size in terms of achieving acceptable confidence limits. My rationale: if you observe a zero effect, the range shouldn't include substantial positive (or beneficial) and substantial negative (or harmful) values. Gives half traditional sample sizes, for 95% confidence limits. But why 95%? 90% can be acceptable and leads to one-third the traditional sample size. The calculations are simple–I won't explain here. This approach is appropriate for studies of mechanisms.

Newest Approach: Clinical Decisiveness You do a study to decide whether an effect is clinically or practically useful or important. You can make two kinds of clinical error with your decision: Type 1: you decide to use an effect that in reality is harmful. Type 2: you decide not to use an effect that in reality is beneficial. You need a big enough sample to keep rates of these errors acceptably low. Acceptably low will depend on how good the benefit is and how bad the harm is. Default: 1% for Type 1 and 20% for Type 2. Leads to sample sizes a bit less than one-third those based on statistical significance.

The Type 1 and 2 error rates are defined by a decision value. If true value = smallest harmful value, and observed value > decision value, you will use the effect in error (rate=1%, say). HARMFULTRIVIALBENEFICIAL Clinical Decisiveness: How It Works, Version 1 If true value = smallest beneficial value, and observed value < decision value, you will not use the effect in error (rate=20%, say). Now solve for the sample size (and the decision value). probability value of effect statistic 0 positivenegative smallest harmful value distribution of observed values, if true value = smallest harmful value distribution of observed values, if true value = smallest beneficial value area = 20% area = 1% decision value smallest beneficial value

How it Works, Version 2 This approach may be easier to understand, because it doesn't involve "if the true value is the smallest worthwhile…". Instead, it's just "worst-case scenario is chances of Type 1 and 2 errors of 1% and 20% (say), which occurs when the observed value is the decision value." Put the observed value on the decision value. Work out the chances that the true effect is harmful and beneficial. You want these to be 1% and 20%. You need to draw a different diagram for this scenario. Solve for the sample size (and the decision value). This approach gives the same answer, of course. Work at it until you understand it!

Conclusions You can justify sample size using adequate precision or acceptable rates of clinical errors. Both make more sense than sample size based on statistical significance and lead to smaller samples. HOWEVER… These sample sizes are for the population mean effect. If there are substantial individual responses, precision or clinical error rates for an individual will be different: Very unlikely may become unlikely or even possible. Your decision for the individual will therefore change. So you need a sample size large enough to characterize individual responses adequately. I'm thinking about it.

This presentation was downloaded from: See Sportscience 10, 2006