7.4 DV’s and Groups Often it is desirous to know if two different groups follow the same or different regression functions -One way to test this is to.

Slides:



Advertisements
Similar presentations
Introduction Describe what panel data is and the reasons for using it in this format Assess the importance of fixed and random effects Examine the Hausman.
Advertisements

Hypothesis Testing Steps in Hypothesis Testing:
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
10.3 Time Series Thus Far Whereas cross sectional data needed 3 assumptions to make OLS unbiased, time series data needs only 2 -Although the third assumption.
Correlation and regression Dr. Ghada Abo-Zaid
3.3 Omitted Variable Bias -When a valid variable is excluded, we UNDERSPECIFY THE MODEL and OLS estimates are biased -Consider the true population model:
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
4.3 Confidence Intervals -Using our CLM assumptions, we can construct CONFIDENCE INTERVALS or CONFIDENCE INTERVAL ESTIMATES of the form: -Given a significance.
8. Heteroskedasticity We have already seen that homoskedasticity exists when the error term’s variance, conditional on all x variables, is constant: Homoskedasticity.
Assumption MLR.3 Notes (No Perfect Collinearity)
8.4 Weighted Least Squares Estimation Before the existence of heteroskedasticity-robust statistics, one needed to know the form of heteroskedasticity -Het.
Choosing a Functional Form
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 13 Additional Topics in Regression Analysis
Multiple Linear Regression Model
Chapter 10 Simple Regression.
Regression with a Binary Dependent Variable. Introduction What determines whether a teenager takes up smoking? What determines if a job applicant is successful.
Additional Topics in Regression Analysis
Economics 20 - Prof. Anderson
4. Multiple Regression Analysis: Estimation -Most econometric regressions are motivated by a question -ie: Do Canadian Heritage commercials have a positive.
7 Dummy Variables Thus far, we have only considered variables with a QUANTITATIVE MEANING -ie: dollars, population, utility, etc. In this chapter we will.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Chapter 11 Multiple Regression.
Topic 3: Regression.
Business Statistics - QBM117 Interval estimation for the slope and y-intercept Hypothesis tests for regression.
EC Prof. Buckles1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 5. Dummy Variables.
1 Javier Aparicio División de Estudios Políticos, CIDE Primavera Regresión.
Relationships Among Variables
Multiple Linear Regression Analysis
12 Autocorrelation Serial Correlation exists when errors are correlated across periods -One source of serial correlation is misspecification of the model.
Lecture 5 Correlation and Regression
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Introduction to Linear Regression and Correlation Analysis
Hypothesis Testing in Linear Regression Analysis
Regression Method.
4.2 One Sided Tests -Before we construct a rule for rejecting H 0, we need to pick an ALTERNATE HYPOTHESIS -an example of a ONE SIDED ALTERNATIVE would.
Quantitative Methods Heteroskedasticity.
1 Research Method Lecture 6 (Ch7) Multiple regression with qualitative variables ©
2-1 MGMG 522 : Session #2 Learning to Use Regression Analysis & The Classical Model (Ch. 3 & 4)
How do Lawyers Set fees?. Learning Objectives 1.Model i.e. “Story” or question 2.Multiple regression review 3.Omitted variables (our first failure of.
CHAPTER 14 MULTIPLE REGRESSION
Chapter 10 Hetero- skedasticity Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
Chap 14-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 14 Additional Topics in Regression Analysis Statistics for Business.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Six.
Ordinary Least Squares Estimation: A Primer Projectseminar Migration and the Labour Market, Meeting May 24, 2012 The linear regression model 1. A brief.
Chapter 4 Linear Regression 1. Introduction Managerial decisions are often based on the relationship between two or more variables. For example, after.
2.4 Units of Measurement and Functional Form -Two important econometric issues are: 1) Changing measurement -When does scaling variables have an effect.
3.4 The Components of the OLS Variances: Multicollinearity We see in (3.51) that the variance of B j hat depends on three factors: σ 2, SST j and R j 2.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
3-1 MGMG 522 : Session #3 Hypothesis Testing (Ch. 5)
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
June 30, 2008Stat Lecture 16 - Regression1 Inference for relationships between variables Statistics Lecture 16.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Example x y We wish to check for a non zero correlation.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
4-1 MGMG 522 : Session #4 Choosing the Independent Variables and a Functional Form (Ch. 6 & 7)
Lecture 6 Feb. 2, 2015 ANNOUNCEMENT: Lab session will go from 4:20-5:20 based on the poll. (The majority indicated that it would not be a problem to chance,
Multiple Regression Analysis with Qualitative Information
Multiple Regression Analysis with Qualitative Information
Multiple Regression Analysis with Qualitative Information
Prepared by Lee Revere and John Large
Tutorial 1: Misspecification
Heteroskedasticity.
Chapter 7: The Normality Assumption and Inference with OLS
Chapter 8: DUMMY VARIABLE (D.V.) REGRESSION MODELS
Multiple Regression Analysis with Qualitative Information
Chapter 9 Dummy Variables Undergraduated Econometrics Page 1
Financial Econometrics Fin. 505
Presentation transcript:

7.4 DV’s and Groups Often it is desirous to know if two different groups follow the same or different regression functions -One way to test this is to use DV’s to allow for ALL different intercepts and slopes for the two groups, and test if all DV terms=0:

7.4 DV’s and Groups As we’ve already seen in our discussion of F tests, the significance or insignificance of one variable is quite separate from the joint significance of a set of variables -Therefore an F test must be done using the restricted model: However, in the case of many x variables, this results in a large number of interaction variables, and may be difficult

7.4 DV’s and Groups Alternately, one can run the regression in question (without DV’s) for the first and second group, and record SSR 1 and SSR 2 -a full regression is then run with all observations included to find SSR p -a test F statistic is then formed as: Which is usually called the CHOW STATISTIC and is only valid under homoskedasticity

7.4 DV’s and Groups This F value is compared to F* from our tables with k+1, n-2(k+1) degrees of freedom -Note that no valid R 2 form of this equation exists -the null hypothesis as listed allows for no difference between groups -if it is not rejected, the two groups test statistically identical (at a certain α) -if one wants to allow for an INTERCEPT difference in the two groups, the full regression is run with a single DV to distinguish the groups

7.5 Binary Dependent Variables -Thus far we have only considered QUANTITATIVE values for our dependent variable, but we can also have dependent variables analyzing a QUALITATIVE event -ie: failing or passing the Midterm -in the simplest case, y has 2 outcomes -ie: MT=1 if passed, =0 otherwise

7.5 Binary Dependent Variables -Here our y value reflect the probability of “success”, that is, -This is also called the RESPONSE PROBABILITY -Since probabilities must sum to one, we also have that -the regression with a binary dependent variable is also called the LINEAR PROBABILITY MODEL (LPM)

7.5 Binary Dependent Variables -in the LPM, B j can no longer be interpretted as the change in y given a one unit change in x. Instead, -Our estimated regression becomes: -Where yhat is our predicted probability of “success” and B j hat predicts the change in the probability of success due to a one unit increase in x j

7.5 Binary Example -Assume that our above example regressed as: -This reflects some LIMITATIONS of the LPM: 1)If bribe (expressed in tens of thousands of dollars)=100 ($1 million), then MThat>4 (400% chance of passing). Ie: estimated probabilities can be negative or over 1. 2)This assumes that the probability increase of the first hour of studying (1%) is the same as the probability increase of the 49 th hour (1%).

7.5 LPM Fixing -One way around this is to redefine predicted values: -One advantage of this redefinition is we can obtain a new goodness-of-fit measure as PERCENT CORRECTLY PREDICTED -as now true and predicted values are both either zero and 1 -the number of matches over the number of observations is our goodness of fit

7.5 LPM and Heteroskedasticity -Because y is binary, the LPM does violate one Gauss-Markov assumption: -where p(X) is short for the probability of success -therefore, heteroskedasticity exists and MLR.5 is violated -therefore t and F statistics are not valid until het is corrected as discussed in chapter 8

7.5 DV Party -It is also possible to include Dummy Variables as both the dependent variable and as independent variables, for example, -Where CheetSheet=1 if a cheat sheet is prepared, =0 otherwise -in this case the estimated coefficient of the independent DV gives the increase in probability of success if not the base case

7.6 Policy Analysis and Program Evaluation -Policy Analysis and Program Evaluation is generally done using the regression and hypothesis test: -Where the DV represents the group possibly needing a program or participation in the program -If H 0 is not rejected, the program is not needed or is not effective

7.6 Evaluation and Analysis Difficulties This evaluation and analysis process has two inherent difficulties: 1)Omitted Variables -if a variable correlated with the DV is omitted, its estimated coefficient and test is invalid -due to past group discrimination, groups are often correlated with factors such as income, education, etc. 2) Self-Selection Problem -often participation in a program is not random

7.6 Omitted Example Consider the following equation: -Where we are testing whether midterm achievement is a function of which class section one is in -However, the choice of class can depend when you eat lunch (among other factors), which can affect Midterm achievement -Therefore by not including eating (which is correlated to our DV), our estimate is biased

7.6 Sample Selection Problems Program evaluation MUST assume that participation in the program or group (and thus inclusion in the control group) is random -however, people often CHOOSE inclusion or non-inclusion (ie: people choose to study, people choose to speed) -since these decisions are influences by parts of the error term, our OLS estimation is biased: Ie: one doesn’t study due to their drug (Survivor) addiction, something that they may not report and thus cannot be included in the regression