2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)

Slides:



Advertisements
Similar presentations
Econometric Modeling Through EViews and EXCEL
Advertisements

Multiple Regression Analysis
Multivariate Regression
Brief introduction on Logistic Regression
Econ 488 Lecture 5 – Hypothesis Testing Cameron Kaplan.
3.3 Omitted Variable Bias -When a valid variable is excluded, we UNDERSPECIFY THE MODEL and OLS estimates are biased -Consider the true population model:
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
(c) 2007 IUPUI SPEA K300 (4392) Outline Least Squares Methods Estimation: Least Squares Interpretation of estimators Properties of OLS estimators Variance.
Bivariate Regression Analysis
Session 2. Applied Regression -- Prof. Juran2 Outline for Session 2 More Simple Regression –Bottom Part of the Output Hypothesis Testing –Significance.
Lecture 4 Econ 488. Ordinary Least Squares (OLS) Objective of OLS  Minimize the sum of squared residuals: where Remember that OLS is not the only possible.
The Simple Linear Regression Model: Specification and Estimation
Multiple Linear Regression Model
CHAPTER 3 ECONOMETRICS x x x x x Chapter 2: Estimating the parameters of a linear regression model. Y i = b 1 + b 2 X i + e i Using OLS Chapter 3: Testing.
Simple Linear Regression
Econ Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Regression Hal Varian 10 April What is regression? History Curve fitting v statistics Correlation and causation Statistical models Gauss-Markov.
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
The Simple Regression Model
FIN357 Li1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
Chapter 11 Multiple Regression.
Topic 3: Regression.
Lecture 2 (Ch3) Multiple linear regression
Simple Linear Regression Analysis
Ordinary Least Squares
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Chapter 11 Simple Regression
Hypothesis Testing in Linear Regression Analysis
Understanding Multivariate Research Berry & Sanders.
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
Specification Error I.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Two Ending Sunday, September 9 (Note: You must go over these slides and complete every.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Six.
OLS SHORTCOMINGS Preview of coming attractions. QUIZ What are the main OLS assumptions? 1.On average right 2.Linear 3.Predicting variables and error term.
1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u.
Properties of OLS How Reliable is OLS?. Learning Objectives 1.Review of the idea that the OLS estimator is a random variable 2.How do we judge the quality.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Ordinary Least Squares Regression.
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
Chapter Three TWO-VARIABLEREGRESSION MODEL: THE PROBLEM OF ESTIMATION
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin The Two-Variable Model: Hypothesis Testing chapter seven.
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
1 Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 1. Estimation.
5. Consistency We cannot always achieve unbiasedness of estimators. -For example, σhat is not an unbiased estimator of σ -It is only consistent -Where.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
10-1 MGMG 522 : Session #10 Simultaneous Equations (Ch. 14 & the Appendix 14.6)
5-1 MGMG 522 : Session #5 Multicollinearity (Ch. 8)
Computacion Inteligente Least-Square Methods for System Identification.
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Regression Overview. Definition The simple linear regression model is given by the linear equation where is the y-intercept for the population data, is.
Ch5 Relaxing the Assumptions of the Classical Model
Kakhramon Yusupov June 15th, :30pm – 3:00pm Session 3
Multivariate Regression
Fundamentals of regression analysis
Two-Variable Regression Model: The Problem of Estimation
Chapter 6: MULTIPLE REGRESSION ANALYSIS
Basic Econometrics Chapter 4: THE NORMALITY ASSUMPTION:
The Regression Model Suppose we wish to estimate the parameters of the following relationship: A common method is to choose parameters to minimise the.
Simple Linear Regression
Tutorial 1: Misspecification
Presentation transcript:

2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)

2-2 Steps in Regression Analysis 1. Review the literature and develop the theoretical model 2. Specify the model: choose X’s and the functional form 3. Hypothesize the expected signs of the coefficients 4. Gather the data 5. Run the regression and evaluate the regression equation 6. Document the results

2-3 Step 1: Review the literature & develop the theoretical model  Do not try desperately to find the regression model that best fits your data.  Rather, you should be able to explain your regression model with some theory.  Therefore, you need to know the relevant theory by reviewing the literature.  If there is no theory directly related to your study, try similar or closely related theory.  If there still is no such closely related theory, you might consult experts in that field or develop your own theory with your story.

2-4 Steps 2-3: Specify the model and hypothesize the signs of the coefficients  The topic you are studying will give you some clues about your dependent variable, Y, to use.  You must decide what X’s to use.  You must decide the functional form: how X’s relate to Y.  The relationships between X’s and Y must be based on some theory, not the data.  The theory that backs up your model should give you some hints about the expected signs of the coefficients in front of the independent variables.

2-5 Steps 4-5: Collect the data, estimate the coefficients, and evaluate  The model should be specified first, before you go out and start collecting data.  Verify the reliability of your data. Look out for errors.  Generally, the more data, the better. [More degrees of freedom]  Unit of measurement does not matter to the regression analysis, except for the interpretation of the coefficients.  Use the data you collected, run the OLS regression to obtain the estimates of the coefficients. You might try other estimators and compare their estimates of the coefficients with those from the OLS.  Then, you must evaluate the estimates of the coefficients. See if the signs are what you expect and significant. See the overall goodness of fit.

2-6 Step 6: Document the results  When documenting the results, the standard format typically is (0.023) (0.023)  n = 100, Adj-R 2 = 0.65  At the minimum, you should include this information.  Some researchers include more information than that given above.  You should describe your data and procedures, and explain your results in words.

2-7 Things to Remember  Do not try to fit a model to your data without some reasonable theory behind your model.  You do not prove the relationships or theory. But you try to find some evidence that there might be some relationships.  Model specification is very important.  Only relevant independent variables should be included in the model.

2-8 The Classical Assumptions 1. The regression model is linear in coefficients, is correctly specified, and has an additive error term. 2. Expected value of the error term is zero. [E( i )=0] 3. All X i ’s are uncorrelated with  i. 4. Error terms are uncorrelated with each other. [No serial correlation] 5. The error term has a constant variance. [No heteroskedasticity] 6. No independent variable can be written as a linear function of other independent variables. [No perfect multicollinearity] 7. The error term is normally distributed. [Optional, but it is nice to have this assumption.]

2-9 Assumption 1: Linear in coefficients, correctly specified, and an additive error term  The model must include all relevant independent variables, must have a correct functional form, and has an additive error term.  The model must be linear in coefficients, but does not have to be linear in variables.

2-10 Assumptions 2-3: E(  i ) = 0, All X i ’s are uncorrelated with  i  The mean of the error term is zero.  The distribution of  i does not have to be a normal distribution, as long as the mean of the distribution is zero.  OLS regression (with an intercept term) will guarantee that the mean of the error term is zero. [By adjusting the intercept if the actual mean is not zero. Therefore, we will pay little attention to the intercept term.]  Assumption 3 states that Cov(X, ) = 0.

2-11 Assumptions 4-5: The error term is uncorrelated with each other, and has a constant variance.  Error terms are uncorrelated with each other. There should not be any information contained in one error term to predict the other error terms.  Very important, especially, in the time- series type of analysis.  If serial correlation exists, it possibly means that you forgot to include some relevant variables in the model.  To get precise estimates of the coefficients, we also need the error term to have a constant variance.

2-12 Assumption 6: No X can be written as a linear function of other X’s.  No perfect multicollinearity such as X 1 = 2X 2 X 1 = X 2 + X 3 X 1 = 5 (a constant)  If there is perfect multicollinearity, you will not be able to obtain the estimates of the coefficients.  Solution: drop one variable from the model.  The existence of perfect multicollinearity is not a problem, but the existence of imperfect multicollinearity is a problem.

2-13 Assumption 7: The error term is normally distributed  Not required for estimation.  But important for hypothesis testing.  The error term can be thought of as the sum of a number of minor influences or errors.  As the number of these minor influences or errors gets large, the distribution of the error term approaches normal distribution [Central Limit Theorem].  Intentionally omitting some variables to achieve normality in the error term should never be considered.

2-14 Sampling Distribution of Estimators  Just as the error term has a distribution, the estimates of the coefficients,, have their own distributions too.  If the error term is normally distributed, the estimates of the coefficients will also be normally distributed because any linear function of a normally distributed variable is itself normally distributed.

2-15 Desirable Properties of the Estimators  We prefer to have an unbiased estimator, the estimator that produces,, where., where.  Among all unbiased estimators to choose from, we also prefer an efficient unbiased estimator, the one that produces the smallest variance of the estimates of the coefficients,.

2-16 Gauss-Markov Theorem & OLS If Assumptions 1-6 are met,  OLS estimator is the minimum variance estimator of all linear unbiased estimators of  k, k = 0, 1, 2,…, K.  “OLS is BLUE.” [Best Linear Unbiased Estimator] –“Best” means that each estimate of  k has the smallest variance possible.

2-17 If Assumptions 1-7 are met,  OLS estimator is the minimum variance estimator of all unbiased estimators of  k, k = 0, 1, 2,…, K.  “OLS is BUE.” [Best Unbiased Estimator]  OLS estimator is equivalent to the Maximum Likelihood (ML) estimator. That is, coefficients estimated from OLS are identical to those estimated from ML.  However, ML gives biased estimate of the variance of the error term, while OLS does not.

2-18 When All Seven Assumptions Are Met OLS estimator 1. is unbiased, 2. has the minimum variance among all estimators, 3. is consistent, and 4. is normally distributed,