Chapter 11: Linear Regression and Correlation

Slides:



Advertisements
Similar presentations
Chapter 12 Simple Linear Regression
Advertisements

Simple Linear Regression
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
Correlation and Simple Regression Introduction to Business Statistics, 5e Kvanli/Guynes/Pavur (c)2000 South-Western College Publishing.
Chapter Topics Types of Regression Models
Correlation and Regression Analysis
Correlation Coefficients Pearson’s Product Moment Correlation Coefficient  interval or ratio data only What about ordinal data?
Correlation & Regression
Objectives of Multiple Regression
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Introduction to Linear Regression and Correlation Analysis
Chapter 11 Simple Regression
Chapter 14 – Correlation and Simple Regression Math 22 Introductory Statistics.
Ms. Khatijahhusna Abd Rani School of Electrical System Engineering Sem II 2014/2015.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Introduction to Probability and Statistics Thirteenth Edition Chapter 12 Linear Regression and Correlation.
1 Multiple Regression A single numerical response variable, Y. Multiple numerical explanatory variables, X 1, X 2,…, X k.
Chapter 11 Correlation and Simple Linear Regression Statistics for Business (Econ) 1.
ECON 338/ENVR 305 CLICKER QUESTIONS Statistics – Question Set #8 (from Chapter 10)
Psychology 202a Advanced Psychological Statistics October 22, 2015.
Chapter 11: Linear Regression and Correlation Regression analysis is a statistical tool that utilizes the relation between two or more quantitative variables.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Chapter 12: Correlation and Linear Regression 1.
Introduction. We want to see if there is any relationship between the results on exams and the amount of hours used for studies. Person ABCDEFGHIJ Hours/
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS St. Edward’s University.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Inference about the slope parameter and correlation
Chapter 13 Simple Linear Regression
The simple linear regression model and parameter estimation
Lecture 11: Simple Linear Regression
Chapter 20 Linear and Multiple Regression
Regression Analysis AGEC 784.
Lecture #26 Thursday, November 17, 2016 Textbook: 14.1 and 14.3
AP Statistics Chapter 14 Section 1.
Correlation and Simple Linear Regression
Virtual COMSATS Inferential Statistics Lecture-26
Linear Regression and Correlation Analysis
Ch12.1 Simple Linear Regression
Chapter 11 Simple Regression
ECONOMETRICS DR. DEEPTI.
Chapter 14: Correlation and Regression
Quantitative Methods Simple Regression.
BIVARIATE REGRESSION AND CORRELATION
Simple Linear Regression - Introduction
Probability and Statistics for Computer Scientists Second Edition, By: Michael Baron Section 11.1: Least squares estimation CIS Computational.
Correlation and Simple Linear Regression
Residuals The residuals are estimate of the error
Regression Models - Introduction
Multiple Regression Models
Lecture Notes The Relation between Two Variables Q Q
Simple Linear Regression
Statistical Assumptions for SLR
Correlation and Simple Linear Regression
The Simple Linear Regression Model: Specification and Estimation
Simple Linear Regression
Simple Linear Regression and Correlation
Simple Linear Regression
Linear Regression and Correlation
Simple Linear Regression
Linear Regression and Correlation
Algebra Review The equation of a straight line y = mx + b
Correlation & Trend Lines
Introduction to Regression
Regression Models - Introduction
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Chapter 11: Linear Regression and Correlation Regression analysis is a statistical tool that utilizes the relation between two or more quantitative variables so that one variable can be predicted from the other, or others. Some Examples: Height and weight of people Income and expenses of people Production size and production time Soil pH and the rate of growth of plants

Correlation An easy way to determine if two quantitative variables are linearly related is by looking at their scatterplot. Another way is to calculate the correlation coefficient, denoted usually by r. The Linear Correlation measures the strength of the linear relationship between explanatory variable (x) and the response variable (y). An estimate of this correlation parameter is provided by the Pearson sample correlation coefficient, r. Note: -1 ≤ r ≤ 1.

Example Scatterplots with Correlations If X and Y are independent, then their correlation is 0.

Correlation If the correlation between X and Y is 0, it doesn’t mean they are independent. It only means that they are not linearly related. Some Guidelines in Interpreting r. One complain about the correlation is that it can be subjective when interpreting its value. Some people are very happy with r≈0.6, while others are not. Value of |r| Strength of linear relationship If |r|≥ 0.95 Very Strong If 0.85 ≤ |r|< 0.95 Strong If 0.65 ≤ |r|< 0.85 Moderate to Strong If 0.45 ≤ |r|< 0.65 Moderate If 0.25 ≤ |r|< 0.45 Weak If|r|< 0.25 Very weak/Close to none Note: Correlation does not necessarily imply Causation!

Simple Linear Regression Model: Yi=(β0+β1xi) + εi where, Yi is the ith value of the response variable. xi is the ith value of the explanatory variable. εi’s are uncorrelated with a mean of 0 and constant variance σ2. Y=β0+β1x Random Error Observed point * y * * * * ε1 * * How do we determine the underlying linear relationship? Well, since the points are following this linear trend, why don’t we look for a line that “best” fit the points. But what do we mean by “best” fit? We need a criterion to help us determine which between 2 competing candidate lines is better. L4 * Expected point L3 x1 x L2 L1

Method of Least Squares Model: Yi=(β0+β1xi) + εi where, Yi is the ith value of the response variable. xi is the ith value of the explanatory variable. εi’s are uncorrelated with a mean of 0 and constant variance σ2. Y=β0+β1x y * Observed * P1(x1,y1) Predicted Example: 2+.8x * y1 * * e1 * * Residual = (Observed y-value) – (Predicted y-value) e1= y1 – e2 * x1 x Method of Least Squares: Choose the line that minimizes the SSE as the “best” line. This line is unknown as the Least-Squares Regression Line. Question: But there are infinite possible candidate lines, how can we find the one that minimizes the SSE? Answer: Since SSE is a continuous function of 2 variables, we can use methods from calculus to minimize the SSE.

Model Assumptions Model: Yi=(β0+β1xi) + εi where, εi’s are uncorrelated with a mean of 0 and constant variance σ2ε. εi’s are normally distributed. (This is needed in the test for the slope.) Y=β0+β1x Since the underlying (green) line is unknown to us, we can’t calculate the values of the error terms (εi). The best that we can do is study the residuals (ei). Observed point y * ε1 e1 Expected point Predicted point x1 x