If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.

Slides:



Advertisements
Similar presentations
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Advertisements

If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Quantitative Data Analysis
A Spreadsheet for Analysis of Straightforward Controlled Trials
Research Skills Workshop Designing a Project
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Client Assessment and Other New Uses of Reliability Will G Hopkins Physiology and Physical Education University of Otago, Dunedin NZ Reliability: the Essentials.
Statistical vs Clinical or Practical Significance
Statistical vs Clinical Significance
Validity and Reliability
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Fitness Chances selected (%) a b d c strength post pre Age Activity r = 0.57 Linear Models and Effect Magnitudes for Research, Clinical and Practical.
Meta-Analyses: Appropriate Growth Will G Hopkins Faculty of Health Science AUT University, Auckland, NZ Resources: Cochrane Reviewers Handbook (2006) at.
Statistical vs. Practical Significance
Understanding p-values Annie Herbert Medical Statistician Research and Development Support Unit
CORRELATION. Overview of Correlation u What is a Correlation? u Correlation Coefficients u Coefficient of Determination u Test for Significance u Correlation.
Logistic Regression Psy 524 Ainsworth.
Statistical Analysis and Data Interpretation What is significant for the athlete, the statistician and team doctor? important Will Hopkins
Correlation Chapter 6. Assumptions for Pearson r X and Y should be interval or ratio. X and Y should be normally distributed. Each X should be independent.
Issues About Statistical Inference Dr R.M. Pandey Additional Professor Department of Biostatistics All-India Institute of Medical Sciences New Delhi.
MAGNITUDE-BASED INFERENCES
Chapter 8 Linear Regression © 2010 Pearson Education 1.
Chapter 11 Multiple Regression.
Notes on Logistic Regression STAT 4330/8330. Introduction Previously, you learned about odds ratios (OR’s). We now transition and begin discussion of.
Social Research Methods
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Multiple Regression – Basic Relationships
Richard M. Jacobs, OSA, Ph.D.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Basic Relationships Purpose of multiple regression Different types of multiple regression.
Correlation Question 1 This question asks you to use the Pearson correlation coefficient to measure the association between [educ4] and [empstat]. However,
Review Regression and Pearson’s R SPSS Demo
Linear Regression and Correlation Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and the level of.
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 12: Multiple and Logistic Regression Marshall University.
Selecting the Correct Statistical Test
Inference for regression - Simple linear regression
Overview of Meta-Analytic Data Analysis
Comparing Means From Two Sets of Data
Introduction To Biological Research. Step-by-step analysis of biological data The statistical analysis of a biological experiment may be broken down into.
Education Research 250:205 Writing Chapter 3. Objectives Subjects Instrumentation Procedures Experimental Design Statistical Analysis  Displaying data.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Correlation and Linear Regression. Evaluating Relations Between Interval Level Variables Up to now you have learned to evaluate differences between the.
POSC 202A: Lecture 12/10 Announcements: “Lab” Tomorrow; Final ed out tomorrow or Friday. I will make it due Wed, 5pm. Aren’t I tender? Lecture: Substantive.
 If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Making Inferences about Effects of Hypoxia/Altitude Will Hopkins sportsci.org/will) Victoria University, Melbourne, Australia  Making.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Contingency tables Brian Healy, PhD. Types of analysis-independent samples OutcomeExplanatoryAnalysis ContinuousDichotomous t-test, Wilcoxon test ContinuousCategorical.
SW388R6 Data Analysis and Computers I Slide 1 Multiple Regression Key Points about Multiple Regression Sample Homework Problem Solving the Problem with.
Testing hypotheses Continuous variables. H H H H H L H L L L L L H H L H L H H L High Murder Low Murder Low Income 31 High Income 24 High Murder Low Murder.
Going from data to analysis Dr. Nancy Mayo. Getting it right Research is about getting the right answer, not just an answer An answer is easy The right.
Statistical inference Statistical inference Its application for health science research Bandit Thinkhamrop, Ph.D.(Statistics) Department of Biostatistics.
(Slides not created solely by me – the internet is a wonderful tool) SW388R7 Data Analysis & Compute rs II Slide 1.
Dr.Rehab F.M. Gwada. Measures of Central Tendency the average or a typical, middle observed value of a variable in a data set. There are three commonly.
Direct method of standardization of indices. Average Values n Mean:  the average of the data  sensitive to outlying data n Median:  the middle of the.
BUS 308 Entire Course (Ash Course) For more course tutorials visit BUS 308 Week 1 Assignment Problems 1.2, 1.17, 3.3 & 3.22 BUS 308.
Making Inferences About Effects Seminar presented at Leeds Beckett and Split universities, March 2016 This slideshow consists of part of the lecture on.
Statistical Analysis and Data Interpretation: What is Important for the Athlete and Statistician Will G Hopkins Institute of Sport and Recreation Research.
Stats Methods at IC Lecture 3: Regression.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
Logistic Regression APKC – STATS AFAC (2016).
Multiple Regression.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
POSC 202A: Lecture Lecture: Substantive Significance, Relationship between Variables 1.
Statistical Analysis and Data Interpretation What is significant for the athlete, the statistician and team doctor? important in health, exercise and sport.
If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it.
15.1 The Role of Statistics in the Research Process
Quantitative Data Analysis
Important Effect Sizes for Exercise and Sport Practitioners and Scientists Will Hopkins sportsci.org/will.
Presentation transcript:

If you are viewing this slideshow within a browser window, select File/Save as… from the toolbar and save the slideshow to your computer, then open it directly in PowerPoint. When you open the file, use the full-screen view to see the information on each slide build sequentially. For full-screen view, click on this icon at the lower left of your screen. To go forwards, left-click or hit the space bar, PdDn or  key. To go backwards, hit the PgUp or  key. To exit from full-screen view, hit the Esc (escape) key.

Magnitude Matters: Effect Size in Research and Clinical Practice Will G Hopkins AUT University, Auckland, NZ Why Magnitude Matters in Research Why Magnitude Matters in Clinical Practice Magnitudes of Effects Types of variables and models Difference between means "Slope" Correlation Difference of proportions Number needed to treat Risk, odds and hazard ratio Difference in mean time to event Age Activity r = 0.57 strength post pre 100 Time males females = innovation! t2 t1 Proportion injured (%)

International Committee of Medical Journal Editors (icmje.org) Background International Committee of Medical Journal Editors (icmje.org) "Show specific effect sizes." "Avoid relying solely on statistical hypothesis testing…, which fails to convey important information about effect size." Publication Manual of the American Psychological Association A section on "Effect Size and Strength of Relationship" 15 ways to express magnitudes. Meta-analysis Emphasis on deriving average magnitude of an effect.

Why Magnitude Matters in Research Two reasons: estimating sample size, and making inferences. Estimating Sample Size Research in our disciplines is all about effects. An effect = a relationship between a predictor variable and a dependent variable. Example: the effect of exercise on a measure of health. We want to know about the effect in a population. But we study a sample of the population. And the magnitude of an effect varies from sample to sample. For a big enough sample, the variation is acceptably small. How many is big enough? Get via statistical, clinical/practical or mechanistic significance. You need the smallest important magnitude of the effect. See MSSE 38(5), 2006: Abstract 2746.

An inference is a statement about the effect in the population. Making Inferences An inference is a statement about the effect in the population. Old approach: is the effect real (statistically significant)? If it isn't, you apparently assume there is no effect. Problem: no mention of magnitude, so depending on sample size… A "real effect" could be clinically trivial. "No effect" could be a clinically clear and useful effect. New approach: is the effect clear? It's clear if it can't be substantially positive and negative. That is, if the confidence interval doesn't overlap such values. New approach: what are the chances the real effect is important? …in a clinical, practical or mechanistic sense. Both new approaches need the smallest important magnitude. You should also make inferences about other magnitudes: small, moderate, large, very large, awe-inspiring.

Why Magnitude Matters in Clinical Practice What really matters is cost-benefit. Here I am addressing only the benefit (and harm). So need smallest important beneficial and harmful magnitudes. Also known as minimum clinically important difference. "A crock"? In the absence of clinical consensus, need statistical defaults. Also need to express in units the clinician, patient, client, athlete, coach or administrator can understand. You should use these terms sometimes: trivial, small, moderate, large, very large, awe-inspiring. The rest of this talk is about these magnitudes for different kinds of effect.

Magnitudes depend on nature of variables. Magnitudes of Effects Magnitudes depend on nature of variables. Continuous: mass, distance, time, current; measures derived therefrom, such as force, concentration, voltage. Counts: such as number of injuries in a season. Nominal: values are levels representing names, such as injured (no, yes), and type of sport (baseball, football, hockey). Ordinal: values are levels with a sense of rank order, such as a 4-pt Likert scale for injury severity (none, mild, moderate, severe). Continuous, counts, ordinals can be treated as numerics, but… As dependents, counts need generalized linear modeling. If ordinal has only a few levels or subjects are stacked at one end, analyze as nominal. Nominals with >2 levels are best dichotomized by comparing or combining levels appropriately. Hard to define magnitude when comparing >2 levels at once.

Effects for linear models: Magnitude also depends on the relationship you model between the dependent and predictor. The model is almost always linear or can be made so. Linear model: sum of predictors and/or their products, plus error. Well developed procedures for estimating effects in linear models. Effects for linear models: Statistical model Effect Predictor Dependent difference in means nominal numeric regression; general linear; mixed; generalized linear Strength Trial "slope" (difference per unit of predictor); correlation numeric Activity Age diffs or ratios of proportions, odds, rates, mean event time nominal logistic regression; generalized linear; proportional hazards InjuredNY Sex "slope" (difference or ratio per unit of predictor) numeric nominal SelectedNY Fitness

Otherwise use the standardized difference or change. Dependent Predictor Effect numeric nominal difference or change in means Strength Trial You consider the difference or change in the mean for pairwise comparisons of levels of the predictor. Clinical or practical experience may give smallest important effect in raw or percent units. Otherwise use the standardized difference or change. Also known as Cohen’s effect size or Cohen's d statistic. You express the difference or change in the mean as a fraction of the between-subject standard deviation (mean/SD). For many measures use the log-transformed dependent variable. It's biased high for small sample size. Correction factor is 1-3/(4v-1), where v=deg. freedom for the SD. The smallest important effect is ±0.2. Trial Strength pre post

Measures of Athletic Performance For team-sport athletes, use standardized differences in mean to get smallest important and other magnitudes. For solo athletes, smallest important effect is 0.3 of a top athlete's typical event-to-event variability. Example: if the variability is a coefficient of variation of 1%, the smallest important effect is 0.3%. This effect would result in a top athlete winning a medal in an extra one competition in 10. I regard moderate, large, very large and extremely large effects as resulting in an extra 3,5, 7 and 9 medals in 10 competitions. Simulation produces the following scale: Note that in many publications I have mistakenly referred to 0.5 of the variability as the smallest effect. <0.3 trivial 0.3-0.9 small 0.9-1.6 moderate 2.5-4.0 very large >4.0 awesome 1.6-2.5 large

Beware: smallest effect on athletic performance depends on how it's measured, because… A percent change in an athlete's ability to output power results in different percent changes in performance in different tests. These differences are due to the power-duration relationship for performance and the power-speed relationship for different modes of exercise. Example: a 1% change in endurance power output produces the following changes… 1% in running time-trial speed or time; ~0.4% in road-cycling time-trial time; 0.3% in rowing-ergometer time-trial time; ~15% in time to exhaustion in a constant-power test. An indeterminable change in any test following a pre-load.

A slope is more practical than a correlation. Dependent Predictor Effect numeric numeric "slope" (difference per unit of predictor); correlation Activity Age A slope is more practical than a correlation. But unit of predictor is arbitrary, so it's hard to define smallest effect for a slope. Example: -2% per year may seem trivial, yet -20% per decade may seem large. For consistency with interpretation of correlation, better to express slope as difference per two SDs of predictor. Fits with smallest important effect of 0.2 SD for the dependent. But underestimates magnitude of larger effects. Easier to interpret the correlation, using Cohen's scale. Smallest important correlation is ±0.1. Complete scale: Age Activity r = 0.57 For an explanation, see newstats.org/effectmag.html <0.1 trivial 0.1-0.3 small 0.3-0.5 moderate >0.5 large 0.7-0.9 very large >0.9 awesome 0.5-0.7 large

You can use correlation to assess nominal predictors. For a two-level predictor, the scales match up. For >2 levels, the correlation doesn't apply to an individual. Magnitudes when controlling for something… Control for = hold it equal or constant or adjust for it. Example: the effect of age on activity adjusted for sex. Control for something by adding it to the model as a predictor. Effect of original predictor changes. No problem for a difference in means or a slope. But correlations are a challenge. The correlation is either partial or semi-partial (SPSS: "part"). Partial = effect of the predictor within a virtual subgroup of subjects who all have the same values of the other predictors. Semi-partial = unique effect of the predictor with all subjects. Partial is probably more appropriate for the individual. Confidence limits may be a problem in some stats packages.

Subjects all start off "N", but different proportions end up "Y". Dependent Predictor Effect nominal nominal differences or ratios of proportions, odds, rates; difference in mean event time InjuredNY Sex Subjects all start off "N", but different proportions end up "Y". Risk difference = a - b. Good measure for an individual, but time dependent. Example: a - b = 83% - 50% = 33%, so extra chance of one in three of injury if you are a male. Smallest effect: ±5%? Number needed to treat (NNT) = 100/(a - b). Number of subjects you would have to treat or sample for one subject to have an outcome attributable to the effect. Example: for every 3 people (=100/33), one extra person would be injured if the people were males. Time (months) 100 Proportion injured (%) males a females b NNT <20 is clinically important?

Odds ratio = (a/c)/(b/d). Population attributable fraction = (a - b)*(fraction population exposed). Smallest important effect for policymakers = ? Relative risk = a/b. Good measure for public health, but time dependent. Smallest effect: 1.1 (or 1/1.1). Based on smallest effect of hazard ratio. Corresponds to risk difference of 55 - 50 = 5%. But relative risk = 6.0 for risk difference = 6 - 1 = 5%. So smallest relative risk for individual is hard to define. Odds ratio = (a/c)/(b/d). Used for logistic regression and some case-control designs. Hard to interpret, but it approximates relative risk when a<10% and b<10% (which is often). Can convert exactly to relative risk if know a or b. a b Time (months) 100 Proportion injured (%) males females c d

Hazard or incidence rate ratio = e/f. Hazard = instantaneous risk rate = proportion per infinitesimal of time. Hazard ratio is best statistical measure. Hazard ratio = risk ratio = odds ratio for low risks (short times). Not dependent on time if incident rates are constant. And even if both rates change, often OK to assume their ratio is constant. Basis of proportional hazards modeling. Smallest effect: 1.1 or 1/1.1. This effect would produce a 10% increase or decrease in the workload of a hospital ward, which would impact personnel and budgets. Time (months) 100 Proportion injured (%) males females e f

Difference in mean time to event = t2 - t1. Best measure for individual when events occur gradually. Can standardize with SD of time to event. Therefore can use default standardized thresholds of 0.2, 0.6, 1.2, 2.0. Bonus: if hazard is constant over time, SD of log(time to event) is independent of hazard. Hence this scale for hazard ratios, derived from standardized thresholds applied to time to event: Proportion injured (%) 100 Time males females t2 t1 100 Time (months) males females t2 t1 <1.28 trivial 1.28-2.0 small 2.0-4.5 moderate 4.5-13 large 13-400 very large >400 awesome

Researchers derive and interpret the slope, not a correlation. Dependent Predictor Effect nominal numeric "slope" (difference or ratio per unit of predictor) SelectedNY Fitness Researchers derive and interpret the slope, not a correlation. Has to be modeled as odds ratio per unit of predictor via logistic regression. Example: odds ratio for selection = 8.1 per unit of fitness. Otherwise same issues as for numeric dependent variable. Need to express as effect of 2 SDs of predictor. When controlling for other predictors, interpret effect as "for subjects all with equal values of the other predictors". Fitness 100 Chances selected (%) predicted chances

This presentation was downloaded from: See Sportscience 10, 2006