Statistics 350 Lecture 21. Today Last Day: Tests and partial R 2 Today: Multicollinearity.

Slides:



Advertisements
Similar presentations
Multicollinearity.
Advertisements

Specification Error II
Lecture 25 Multiple Regression Diagnostics (Sections )
Statistics 350 Lecture 16. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
Lecture 24 Multiple Regression (Sections )
Ekonometrika 1 Ekonomi Pembangunan Universitas Brawijaya.
Empirical Estimation Review EconS 451: Lecture # 8 Describe in general terms what we are attempting to solve with empirical estimation. Understand why.
Statistics 350 Lecture 17. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Regression, Residuals, and Coefficient of Determination Section 3.2.
THE DUMMY VARIABLE TRAP 1 Suppose that you have a regression model with Y depending on a set of ordinary variables X 2,..., X k and a qualitative variable.
Objectives of Multiple Regression
Chapter 14 – Correlation and Simple Regression Math 22 Introductory Statistics.
A P STATISTICS LESSON 3 – 2 CORRELATION.
2.4: Cautions about Regression and Correlation. Cautions: Regression & Correlation Correlation measures only linear association. Extrapolation often produces.
Lecture 17 Summary of previous Lecture Eviews. Today discussion  R-Square  Adjusted R- Square  Game of Maximizing Adjusted R- Square  Multiple regression.
STA302/ week 111 Multicollinearity Multicollinearity occurs when explanatory variables are highly correlated, in which case, it is difficult or impossible.
MultiCollinearity. The Nature of the Problem OLS requires that the explanatory variables are independent of error term But they may not always be independent.
Lecture 22 Dustin Lueker.  The sample mean of the difference scores is an estimator for the difference between the population means  We can now use.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
AP Statistics Chapter 8 & 9 Day 3
Soc 3306a Lecture 9: Multivariate 2 More on Multiple Regression: Building a Model and Interpreting Coefficients.
2 Multicollinearity Presented by: Shahram Arsang Isfahan University of Medical Sciences April 2014.
Example 1: page 161 #5 Example 2: page 160 #1 Explanatory Variable - Response Variable - independent variable dependent variable.
Holt Algebra Curve Fitting with Linear Models 2-7 Curve Fitting with Linear Models Holt Algebra 2 Lesson Presentation Lesson Presentation.
1 Everyday is a new beginning in life. Every moment is a time for self vigilance.
© 2013 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Introductory Statistics: Exploring the World through.
CHAPTER 4: Scatterplots and Correlation ESSENTIAL STATISTICS Second Edition David S. Moore, William I. Notz, and Michael A. Fligner Lecture Presentation.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Stat 112 Notes 9 Today: –Multicollinearity (Chapter 4.6) –Multiple regression and causal inference.
May 2004 Prof. Himayatullah 1 Basic Econometrics Chapter 7 MULTIPLE REGRESSION ANALYSIS: The Problem of Estimation.
Click to edit Master title style Midterm 3 Wednesday, June 10, 1:10pm.
AP STATISTICS LESSON 4 – 2 ( DAY 1 ) Cautions About Correlation and Regression.
Regression and Least Squares The need for a mathematical construct… Insert fig 3.8.
LECTURE 9 Tuesday, 24 FEBRUARY STA291 Fall Administrative 4.2 Measures of Variation (Empirical Rule) 4.4 Measures of Linear Relationship Suggested.
Chapter 4 Summary Scatter diagrams of data pairs (x, y) are useful in helping us determine visually if there is any relation between x and y values and,
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
SWBAT: Calculate and interpret the residual plot for a line of regression Do Now: Do heavier cars really use more gasoline? In the following data set,
A P STATISTICS LESSON 3 – 3 (DAY 3) A P STATISTICS LESSON 3 – 3 (DAY 3) RISIDUALS.
I271B QUANTITATIVE METHODS Regression and Diagnostics.
June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16.
Copyright © 2014, 2011 Pearson Education, Inc. 1 Active Learning Lecture Slides For use with Classroom Response Systems Chapter 23 Multiple Regression.
Notes Chapter 7 Bivariate Data. Relationships between two (or more) variables. The response variable measures an outcome of a study. The explanatory variable.
Correlation and Median-Median Line Statistics Test: Oct. 20 (Wednesday)
The coefficient of determination, r 2, is The fraction of the variation in the value of y that is explained by the regression line and the explanatory.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Correlation  We can often see the strength of the relationship between two quantitative variables in a scatterplot, but be careful. The two figures here.
MATH 2311 Section 5.2 & 5.3. Correlation Coefficient.
GOAL: I CAN USE TECHNOLOGY TO COMPUTE AND INTERPRET THE CORRELATION COEFFICIENT OF A LINEAR FIT. (S-ID.8) Data Analysis Correlation Coefficient.
Predicting Energy Consumption in Buildings using Multiple Linear Regression Introduction Linear regression is used to model energy consumption in buildings.
REGRESSION (R2).
REGRESSION DIAGNOSTIC I: MULTICOLLINEARITY
Fundamentals of regression analysis
MULTIPLE REGRESSION ANALYSIS: THE PROBLEM OF ESTIMATION
DRQ 8 Dr. Capps AGEC points
Simple Linear Regression
Tutorial 8 Table 3.10 on Page 76 shows the scores in the final examination F and the scores in two preliminary examinations P1 and P2 for 22 students in.
STA 291 Summer 2008 Lecture 23 Dustin Lueker.
11A Correlation, 11B Measuring Correlation
Performing a regression analysis
y = mx + b Linear Regression line of best fit REMEMBER:
Statistics 350 Lecture 18.
Correlation & Trend Lines
STA 291 Spring 2008 Lecture 23 Dustin Lueker.
Chapter 3 Examining Relationships
Statistics 101 CORRELATION Section 3.2.
Multicollinearity Multicollinearity occurs when explanatory variables are highly correlated, in which case, it is difficult or impossible to measure their.
Solution to Problem 2.25 DS-203 Fall 2007.
3.2 Correlation Pg
Presentation transcript:

Statistics 350 Lecture 21

Today Last Day: Tests and partial R 2 Today: Multicollinearity

Multicollinearity When explanatory variables are highly correlated, weird things can happen with the regression analysis Multicollinearity is said to exist among explanatory variables any time a regression of one of the explanatory variables against the rest yields a strong linear relationship, as measured by a high R 2 Can also attempt to visualize this relationship using a scatter-plot matrix

Multicollinearity Back to Example:

Multicollinearity Why might this matter? Consider the 3 variable linear regression model: Can view  1 in the model as the partial regression coefficient for X 1 What is its interpretation?

Multicollinearity If other variables tend to be correlated with X 1 this effect is difficult to isolate and estimate RESULT:

Multicollinearity Back to example:

Multicollinearity Back to example: Notice: If did a regression of X 1 on X 2 and X 3, the R 2 is Conclusion:

Multicollinearity Why exactly have we observed this phenomenon? Consider the 3 variable model in the body fat example:

Multicollinearity As a result:

Multicollinearity Detecting multicollinearity in practice: