Correlation Coefficients Pearson’s Product Moment Correlation Coefficient  interval or ratio data only What about ordinal data?

Slides:



Advertisements
Similar presentations
Kin 304 Regression Linear Regression Least Sum of Squares
Advertisements

Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Simple Linear Regression. G. Baker, Department of Statistics University of South Carolina; Slide 2 Relationship Between Two Quantitative Variables If.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Simple Linear Regression
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
9. SIMPLE LINEAR REGESSION AND CORRELATION
Chapter Eighteen MEASURES OF ASSOCIATION
Chapter Topics Types of Regression Models
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
REGRESSION AND CORRELATION
Measures of Association Deepak Khazanchi Chapter 18.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Simple Linear Regression and Correlation
Simple Linear Regression Analysis
Statistical hypothesis testing – Inferential statistics II. Testing for associations.
Correlation & Regression
Introduction to Linear Regression and Correlation Analysis
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Correlation and Regression
Chapter 11 Simple Regression
Simple Linear Regression
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Chapter 14 – Correlation and Simple Regression Math 22 Introductory Statistics.
CHAPTER 15 Simple Linear Regression and Correlation
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
Regression. Population Covariance and Correlation.
Examining Relationships in Quantitative Research
PS 225 Lecture 20 Linear Regression Equation and Prediction.
Correlation and Regression Basic Concepts. An Example We can hypothesize that the value of a house increases as its size increases. Said differently,
Chapter 16 Data Analysis: Testing for Associations.
Chapter 6 Simple Regression Introduction Fundamental questions – Is there a relationship between two random variables and how strong is it? – Can.
Simple Linear Regression In the previous lectures, we only focus on one random variable. In many applications, we often work with a pair of variables.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
A Significance Test for r An estimator r    = 0 ? t-test.
Examining Relationships in Quantitative Research
Chapter Thirteen Copyright © 2006 John Wiley & Sons, Inc. Bivariate Correlation and Regression.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Chapter Thirteen Bivariate Correlation and Regression Chapter Thirteen.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Psychology 202a Advanced Psychological Statistics October 22, 2015.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics Seventh Edition By Brase and Brase Prepared by: Lynn Smith.
Chapter 11: Linear Regression and Correlation Regression analysis is a statistical tool that utilizes the relation between two or more quantitative variables.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
SOCW 671 #11 Correlation and Regression. Uses of Correlation To study the strength of a relationship To study the direction of a relationship Scattergrams.
Correlation and Regression Basic Concepts. An Example We can hypothesize that the value of a house increases as its size increases. Said differently,
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Research Methods: 2 M.Sc. Physiotherapy/Podiatry/Pain Correlation and Regression.
Simple Linear Regression In many scientific investigations, one is interested to find how something is related with something else. For example the distance.
The simple linear regression model and parameter estimation
Chapter 11: Linear Regression and Correlation
Chapter 20 Linear and Multiple Regression
Regression and Correlation
REGRESSION G&W p
Kin 304 Regression Linear Regression Least Sum of Squares
BPK 304W Regression Linear Regression Least Sum of Squares
Quantitative Methods Simple Regression.
Correlation and Simple Linear Regression
1) A residual: a) is the amount of variation explained by the LSRL of y on x b) is how much an observed y-value differs from a predicted y-value c) predicts.
Simple Linear Regression
PENGOLAHAN DAN PENYAJIAN
Correlation and Simple Linear Regression
Simple Linear Regression
Simple Linear Regression
Presentation transcript:

Correlation Coefficients Pearson’s Product Moment Correlation Coefficient  interval or ratio data only What about ordinal data?

Spearman’s Rank Correlation Coefficient r s = 1 -  di2di2 i=1 i=n n 3 - n 6

Spearman’s Rank Correlation Coefficient: Example

A Significance Test for r s SE r s = 1 n -1 t test = rsrs SE r s = r s n -1 df = n - 1

Spearman’s Rank Correlation Coefficient: Example

Pearson’s r - Assumptions 1.Interval or ratio scale data 2.Selected randomly 3.Linear 4.Joint bivariate normal distribution  S-Plus (qqnorm)

Spearman’s Rank Correlation Coefficient Ordinal data already in a ranked form Interval or ratio data convert it to rankings

Spearman’s Rank Correlation Coefficient TVDI (x) Rank (x) Theta (y) Rank (y) Difference (d i )

A Significance Test for r s

 S-Plus

TVDI (x) Theta (y)

Correlation  Direction & Strength We might wish to go a little further Rate of change Predictability Correlation  Regression

Deterministic  perfect knowledge Probabilistic  estimate  not with absolute accuracy (or certainty) Two Sorts of Bivariate Relationships

Travel  at a constant speed Deterministic  time spent driving vs. distance traveled A Deterministic Relationship s = s 0 + vt s: distance traveled s 0 : initial distance v: speed t: time traveled time (t) distance (s) slope (v) intercept (s 0 ) Truly deterministic  rare

More often  probabilistic e.g., ages vs. heights (2 – 20 yrs) A Probabilistic Relationship age (years) height (meters) Good relationship Unpredictability or error

Sampling and Regression Our expectation (less than perfect) Collecting data  measurement errors  height Other factors (not accounted for in the model)  plant growth vs. T

Simple vs. Multiple Regression Simple linear regression  y  x Multiple linear regression  y  x 1, x 2, … x n

Model y = a + bx + e Simple Linear Regression x: independent variable y: dependent variable b: slope a: intercept e: error term x (independent) y (dependent) b a error: 

Scatterplot  fitting a line Fitting a Line to a Set of Points x (independent) y (dependent) Least squares method Minimize the error term e

Sampling and Regression Sampled data  model y = a + bx + e Attempt to estimate a “true” regression line y =  +  x +  Multiple samples  several similar regression lines  the population regression line

Minimize the error term e The line of best fit  ŷ = a + b Least Squares Method y ŷ = a + bx ŷ (y - ŷ)

Estimates and Residuals Errors e = y – ŷ Residuals  Underestimate  Overestimate

Errors (residuals)  e = (y - ŷ) Overall error  Simply sum these error terms  0  Square the differences and then sum them up to create a useful estimate Minimizing the Error Term SSE =   (y - ŷ) 2 i = 1 n 

Minimizing the SSE   (y - ŷ) 2 i = 1 n min a,b n   (y i - a - bx i ) 2 i = 1 min a,b =

Least squares method  Finding Regression Coefficients   (x i - x) (y i - y) i = 1 n b =   (x i - x) 2 i = 1 n a = y - bx

Interpreting Slope (b) Slope of the line (b  the change in y due to a unit change in x b > 0 b < 0

Regression Slope and Correlation  (x i - x)(y i - y) i=1 i=n (n - 1) s X s Y r =   (x i - x) (y i - y) i = 1 n b =   (x i - x) 2 i = 1 n b = r sysy sxsx