Correlation Assume you have two measurements, x and y, on a set of objects, and would like to know if x and y are related. If they are directly related,

Slides:



Advertisements
Similar presentations
Introduction to Hypothesis Testing
Advertisements

Hypothesis Testing Steps in Hypothesis Testing:
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
Correlation and Linear Regression
Chapter 13 Multiple Regression
Statistical Significance What is Statistical Significance? What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant?
HYPOTHESIS TESTING Four Steps Statistical Significance Outcomes Sampling Distributions.
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Chapter 12 Multiple Regression
Statistical Significance What is Statistical Significance? How Do We Know Whether a Result is Statistically Significant? How Do We Know Whether a Result.
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
The Simple Regression Model
Chapter Topics Types of Regression Models
Correlation A correlation exists between two variables when one of them is related to the other in some way. A scatterplot is a graph in which the paired.
Chapter 11 Multiple Regression.
Correlation and Regression. Correlation What type of relationship exists between the two variables and is the correlation significant? x y Cigarettes.
REGRESSION AND CORRELATION
Chapter 9: Correlation and Regression
Linear Regression Example Data
Today Concepts underlying inferential statistics
Correlation and Regression Analysis
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Chapter 9: Introduction to the t statistic
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
Chapter 9 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 What is a Perfect Positive Linear Correlation? –It occurs when everyone has the.
Relationships Among Variables
Lecture 5 Correlation and Regression
Lecture 16 Correlation and Coefficient of Correlation
Regression and Correlation Methods Judy Zhong Ph.D.
AM Recitation 2/10/11.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Correlation and Linear Regression
Chapter 8 Introduction to Hypothesis Testing
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
Section #6 November 13 th 2009 Regression. First, Review Scatter Plots A scatter plot (x, y) x y A scatter plot is a graph of the ordered pairs (x, y)
CORRELATION & REGRESSION
Correlation.
Correlation and Regression
Chapter 15 Correlation and Regression
Inferential Statistics 2 Maarten Buis January 11, 2006.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
● Final exam Wednesday, 6/10, 11:30-2:30. ● Bring your own blue books ● Closed book. Calculators and 2-page cheat sheet allowed. No cell phone/computer.
Introduction to Linear Regression
EQT 373 Chapter 3 Simple Linear Regression. EQT 373 Learning Objectives In this chapter, you learn: How to use regression analysis to predict the value.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
Elementary Statistics Correlation and Regression.
Biostatistics Class 6 Hypothesis Testing: One-Sample Inference 2/29/2000.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
Lecture 10: Correlation and Regression Model.
Chapter 14 Correlation and Regression
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Lecture 1: Basic Statistical Tools. A random variable (RV) = outcome (realization) not a set value, but rather drawn from some probability distribution.
Chapter 13 Understanding research results: statistical inference.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Hypothesis Testing and Statistical Significance
Correlation and Regression Elementary Statistics Larson Farber Chapter 9 Hours of Training Accidents.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Regression and Correlation
Chapter 9 Hypothesis Testing.
When You See (This), You Think (That)
Chapter 7: The Normality Assumption and Inference with OLS
Inferential Statistics
Presentation transcript:

Correlation Assume you have two measurements, x and y, on a set of objects, and would like to know if x and y are related. If they are directly related, when x is high, y tends to be high, and when x is low, y tends to be low. If they are indirectly related, when x is high, y tends to be low, and when x is low, y tends to be high. This suggests summing the products of the z- scores.

xy case Scatter plot of y vs. x

x (zscore) y (zscore) case mean xy xzyzxzyz Correlation coefficient (Pearson’s) r

Properties of r r ranges from -1.0 to +1.0 r = 1 means perfect linear relationship r = -1 means perfect linear relationship with negative slope r = 0 means no correlation

r = r = Example scatterplots r = -.24r =.41 r = 0 r = -.66r =.94

Correlation and causation “Correlation does not imply causation” More precisely, x correlated with y does not imply x causes y, because correlation could be a type I error y could cause x z could cause both x and y

Uncorrelated does not mean independent x y x is highly predictive of y, but r = 0

Significance test for r The aim is to test the null hypothesis that the population correlation ρ (rho) is 0. The larger n, the less likely a given r will happen under the null hypothesis. From r and n, we can compute a p-value From n and α, we can compute a critical r Numerical example

Regression Correlation suggests a linear model of y as a function of x A linear model is defined by ŷ = mx + b + e random error with mean 0equation for a line slopeinterceptpredicted y

x regression y residuals x e R 2 = , F = , p = Regression line: y = x

r vs. R 2 R 2 is actually the square of r. So why is it capitalized and squared in a regression? r ranges from -1 to 1. But in a regression, r cannot meaningfully be negative, because it is the correlation between y and ŷ. Since ŷ is the best estimate of y, this correlation is automatically positive. The capitalization and squaring reflects this situation. It is squared to

Interpretation of R2 R 2 can be interpreted as the proportion of the variance accounted for R 2 = 1 - SS error SS total SS reg SS total = regression line mean R 2 is high when the unexplained (residual) variance is small relative to the total amount of variance

Simpson’s paradox Size of animal length of ears Negatively correlated Or positively correlated? Rabbits Humans Whales Adding a variable can change the sign of the correlation

Effect size Beyond computing significance, we often need an estimate of the magnitude of an effect. There are two basic ways of expressing this: - Normalized mean difference - Proportion of variance accounted for

The normalized difference between means Cohen’s d expresses how the difference between two means relative to the spread in the data.

Proportion of variance accounted for R 2 can be interpreted as the proportion of all the variance in the data that is predicted by a regression model η 2 (eta squared) can be interpreted as the proportion of all variance in a factorial design that is accounted for by main effects and interactions

Power Power is the probability of finding an effect of a given size in your experiment, i.e. The probability of rejecting the null hypothesis if the null hypothesis is actually false.

Outliers An outlier is a measurement that is so discrepant from others that it seems “suspicious.” If p(x suspicious |distribution) is low enough, we “reject the null hypothesis” that x suspicious came from the same distribution as the others, and remove it. A common rule of thumb is z > 2.5 (or 2 or 3), BUT... But also consider transforms that avoid outliers in the first place, like 1/x. Removed data is best NOT REPLACED. But if it must be replaced, do so “conservatively,” i.e. in a manner biased towards the null hypothesis.

Chi squared Assume that mutually K exclusive outcomes are predicted to occur E 1,E 2,...,E K, times...but are actually observed to occur N 1,N 2,...,N K times respectively... A chi-square test allows us to evaluate the null hypothesis that the proportions were as expected, with deviations “by chance.”

Performing a chi-squared test For each outcome, compute Sum them up over all outcomes Then, under the null hypothesis, this total will be distributed as a χ 2 distribution with n-1 degrees of freedom.

The Bayesian perspective Conventional statistics is based on a frequentist definition of probability, which insists that hypotheses do not have “probabilities.” → All we can do is “reject” H, or not reject it. Bayesian inference is based on a subjectivist definition of probability, which considers p(H) to be the “degree of belief” in hypothesis H, simply expressing our uncertainty about H in light of the data. → Instead of accepting or rejecting, we seek p(H|E).

Cartoon 1: Fisher Fisher: Given the sampling distribution of the null p(E|H 0 ), consider the likelihood of the null hypothesis, integrated out to the tail. If this probability is low, this tends to contradict the null hypothesis. In fact, if it is lower than.05, we informally “reject” the null. 0 p(E|H0)p(E|H0) E probability density

Cartoon 2: Neyman & Pearson N&P: There are really two hypotheses, the null H 0 and some alternative H 1. Our main goal is to avoid a Type I error. So set this probability at α, which determines our criterion for rejecting the null. Note though that there is also a possibility of making a Type II error, a hit, or a correct rejection. Compute power and set sample size to control the probability of a Type II error. 0 p(E|H0)p(E|H0) Expected effect size p(E|H1)p(E|H1) μ1μ1 E probability density

Cartoon 3: Bayes (/Laplace/Jeffreys) What we really want is to evaluate how strongly our data favors either hypothesis, not just make an accept/reject decision. For each H, the degree of belief in it, conditioned on the data, is p(H|E). So to evaluate the relative strength of the H 1 and H 0, consider the posterior ratio This expresses how strongly the data and priors favor H 1 relative to H 0, taking into account everything we know about the situation. Degree of belief in H 1 Degree of belief in H 0 =

Decomposing the posterior ratio posterior ratio = prior ratio × likelihood ratio If you want to be “unbiased”, set the prior ratio to 1, sometimes called an “uninformative prior.” Then your posterior belief about H 0 and H 1 depends entirely on the likelihood ratio, aka “Bayes factor.”

Visualizing the Likelihood Ratio 0 Expected effect size μ1μ1 = p(E|H0)p(E|H0)p(E|H1)p(E|H1) height of green bar at E height of red bar at E E probability density

Interpretation of likelihood ratios LR = 1 means the evidence was neutral about which hypothesis was correct. LR > 1 means the evidence favors the hypothesis. Jeffreys (1939) suggested rules of thumb, e.g. LR > 3 means “substantial” evidence in favor of H 1, LR >10 means “strong,” evidence etc. LR < 1 means the evidence actually favored the null hypothesis.

LRs vs. p-values Likelihood ratios and p- values are not at all the same thing. But in practice, they are related. Dixon (1998)