CHOW TEST AND DUMMY VARIABLE GROUP TEST

Slides:



Advertisements
Similar presentations
1 Although they are biased in finite samples if Part (2) of Assumption C.7 is violated, OLS estimators are consistent if Part (1) is valid. We will demonstrate.
Advertisements

ADAPTIVE EXPECTATIONS 1 The dynamics in the partial adjustment model are attributable to inertia, the drag of the past. Another, completely opposite, source.
EXPECTED VALUE RULES 1. This sequence states the rules for manipulating expected values. First, the additive rule. The expected value of the sum of two.
ADAPTIVE EXPECTATIONS: FRIEDMAN'S PERMANENT INCOME HYPOTHESIS
EC220 - Introduction to econometrics (chapter 14)
EC220 - Introduction to econometrics (chapter 11)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 9) Slideshow: two-stage least squares Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 12) Slideshow: consequences of autocorrelation Original citation: Dougherty, C. (2012)
EC220 - Introduction to econometrics (chapter 1)
THE ERROR CORRECTION MODEL 1 The error correction model is a variant of the partial adjustment model. As with the partial adjustment model, we assume a.
1 MAXIMUM LIKELIHOOD ESTIMATION OF REGRESSION COEFFICIENTS X Y XiXi 11  1  +  2 X i Y =  1  +  2 X We will now apply the maximum likelihood principle.
EC220 - Introduction to econometrics (chapter 3)
1 This very short sequence presents an important definition, that of the independence of two random variables. Two random variables X and Y are said to.
EC220 - Introduction to econometrics (review chapter)
EC220 - Introduction to econometrics (chapter 5)
EC220 - Introduction to econometrics (chapter 10)
EC220 - Introduction to econometrics (chapter 4)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: slope dummy variables Original citation: Dougherty, C. (2012) EC220 -
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: interactive explanatory variables Original citation: Dougherty, C. (2012)
HETEROSCEDASTICITY-CONSISTENT STANDARD ERRORS 1 Heteroscedasticity causes OLS standard errors to be biased is finite samples. However it can be demonstrated.
EC220 - Introduction to econometrics (chapter 7)
EC220 - Introduction to econometrics (chapter 2)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 6) Slideshow: variable misspecification iii: consequences for diagnostics Original.
EC220 - Introduction to econometrics (chapter 1)
1 INTERPRETATION OF A REGRESSION EQUATION The scatter diagram shows hourly earnings in 2002 plotted against years of schooling, defined as highest grade.
TESTING A HYPOTHESIS RELATING TO A REGRESSION COEFFICIENT This sequence describes the testing of a hypotheses relating to regression coefficients. It is.
SLOPE DUMMY VARIABLES 1 The scatter diagram shows the data for the 74 schools in Shanghai and the cost functions derived from a regression of COST on N.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 3) Slideshow: precision of the multiple regression coefficients Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: semilogarithmic models Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 4) Slideshow: nonlinear regression Original citation: Dougherty, C. (2012) EC220 -
DERIVING LINEAR REGRESSION COEFFICIENTS
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: Chow test Original citation: Dougherty, C. (2012) EC220 - Introduction.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: dummy variable classification with two categories Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: two sets of dummy variables Original citation: Dougherty, C. (2012) EC220.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: the effects of changing the reference category Original citation: Dougherty,
Christopher Dougherty EC220 - Introduction to econometrics (chapter 5) Slideshow: dummy classification with more than two categories Original citation:
DUMMY CLASSIFICATION WITH MORE THAN TWO CATEGORIES This sequence explains how to extend the dummy variable technique to handle a qualitative explanatory.
1 INTERACTIVE EXPLANATORY VARIABLES The model shown above is linear in parameters and it may be fitted using straightforward OLS, provided that the regression.
1 TWO SETS OF DUMMY VARIABLES The explanatory variables in a regression model may include multiple sets of dummy variables. This sequence provides an example.
Confidence intervals were treated at length in the Review chapter and their application to regression analysis presents no problems. We will not repeat.
1 PROXY VARIABLES Suppose that a variable Y is hypothesized to depend on a set of explanatory variables X 2,..., X k as shown above, and suppose that for.
MULTIPLE RESTRICTIONS AND ZERO RESTRICTIONS
F TEST OF GOODNESS OF FIT FOR THE WHOLE EQUATION 1 This sequence describes two F tests of goodness of fit in a multiple regression model. The first relates.
MULTIPLE REGRESSION WITH TWO EXPLANATORY VARIABLES: EXAMPLE 1 This sequence provides a geometrical interpretation of a multiple regression model with two.
Simple regression model: Y =  1 +  2 X + u 1 We have seen that the regression coefficients b 1 and b 2 are random variables. They provide point estimates.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 9) Slideshow: instrumental variable estimation: variation Original citation: Dougherty,
. reg LGEARN S WEIGHT85 Source | SS df MS Number of obs = F( 2, 537) = Model |
Christopher Dougherty EC220 - Introduction to econometrics (chapter 6) Slideshow: multiple restrictions and zero restrictions Original citation: Dougherty,
Chapter 5: Dummy Variables. DUMMY VARIABLE CLASSIFICATION WITH TWO CATEGORIES 1 We’ll now examine how you can include qualitative explanatory variables.
(1)Combine the correlated variables. 1 In this sequence, we look at four possible indirect methods for alleviating a problem of multicollinearity. POSSIBLE.
1 We will continue with a variation on the basic model. We will now hypothesize that p is a function of m, the rate of growth of the money supply, as well.
COST 11 DUMMY VARIABLE CLASSIFICATION WITH TWO CATEGORIES 1 This sequence explains how you can include qualitative explanatory variables in your regression.
RAMSEY’S RESET TEST OF FUNCTIONAL MISSPECIFICATION 1 Ramsey’s RESET test of functional misspecification is intended to provide a simple indicator of evidence.
1 NONLINEAR REGRESSION Suppose you believe that a variable Y depends on a variable X according to the relationship shown and you wish to obtain estimates.
1 ESTIMATORS OF VARIANCE, COVARIANCE, AND CORRELATION We have seen that the variance of a random variable X is given by the expression above. Variance.
1 CHANGES IN THE UNITS OF MEASUREMENT Suppose that the units of measurement of Y or X are changed. How will this affect the regression results? Intuitively,
SEMILOGARITHMIC MODELS 1 This sequence introduces the semilogarithmic model and shows how it may be applied to an earnings function. The dependent variable.
GRAPHING A RELATIONSHIP IN A MULTIPLE REGRESSION MODEL The output above shows the result of regressing EARNINGS, hourly earnings in dollars, on S, years.
1 REPARAMETERIZATION OF A MODEL AND t TEST OF A LINEAR RESTRICTION Linear restrictions can also be tested using a t test. This involves the reparameterization.
F TESTS RELATING TO GROUPS OF EXPLANATORY VARIABLES 1 We now come to more general F tests of goodness of fit. This is a test of the joint explanatory power.
WHITE TEST FOR HETEROSCEDASTICITY 1 The White test for heteroscedasticity looks for evidence of an association between the variance of the disturbance.
1 COMPARING LINEAR AND LOGARITHMIC SPECIFICATIONS When alternative specifications of a regression model have the same dependent variable, R 2 can be used.
VARIABLE MISSPECIFICATION II: INCLUSION OF AN IRRELEVANT VARIABLE In this sequence we will investigate the consequences of including an irrelevant variable.
FOOTNOTE: THE COCHRANE–ORCUTT ITERATIVE PROCESS 1 We saw in the previous sequence that AR(1) autocorrelation could be eliminated by a simple manipulation.
VARIABLE MISSPECIFICATION I: OMISSION OF A RELEVANT VARIABLE In this sequence and the next we will investigate the consequences of misspecifying the regression.
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Introduction to Econometrics, 5th edition
Presentation transcript:

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST N In the dummy variable sequences and in the Chow test sequence we investigated whether the cost functions for occupational and regular schools are different. 1

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST N In each case we performed tests that showed that the functions are significantly different. Could the two approaches have led to different conclusions? 2

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST N The answer is no. The Chow test is equivalent to an F test testing the explanatory power of the dummy variables as a group. 3

CHOW TEST AND DUMMY VARIABLE GROUP TEST . reg COST N Source | SS df MS Number of obs = 74 ---------+------------------------------ F( 1, 72) = 46.82 Model | 5.7974e+11 1 5.7974e+11 Prob > F = 0.0000 Residual | 8.9160e+11 72 1.2383e+10 R-squared = 0.3940 ---------+------------------------------ Adj R-squared = 0.3856 Total | 1.4713e+12 73 2.0155e+10 Root MSE = 1.1e+05 ------------------------------------------------------------------------------ COST | Coef. Std. Err. t P>|t| [95% Conf. Interval] ---------+-------------------------------------------------------------------- N | 339.0432 49.55144 6.842 0.000 240.2642 437.8222 _cons | 23953.3 27167.96 0.882 0.381 -30205.04 78111.65 With both approaches the starting point is a simple regression of annual recurrent expenditure on the number of students enrolled, using the entire sample. We make a note of RSS. 4

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST N The regression line is shown graphically. 5

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST occupational school regular school N We now make a distinction between occupational schools and regular schools. 6

CHOW TEST AND DUMMY VARIABLE GROUP TEST . reg COST N OCC NOCC Source | SS df MS Number of obs = 74 ---------+------------------------------ F( 3, 70) = 49.64 Model | 1.0009e+12 3 3.3363e+11 Prob > F = 0.0000 Residual | 4.7045e+11 70 6.7207e+09 R-squared = 0.6803 ---------+------------------------------ Adj R-squared = 0.6666 Total | 1.4713e+12 73 2.0155e+10 Root MSE = 81980 ------------------------------------------------------------------------------ COST | Coef. Std. Err. t P>|t| [95% Conf. Interval] ---------+-------------------------------------------------------------------- N | 152.2982 60.01932 2.537 0.013 32.59349 272.003 OCC | -3501.177 41085.46 -0.085 0.932 -85443.55 78441.19 NOCC | 284.4786 75.63211 3.761 0.000 133.6351 435.3221 _cons | 51475.25 31314.84 1.644 0.105 -10980.24 113930.7 With the dummy variable approach, we take one type of school as the reference category. We will choose regular schools for this category, but it makes no difference to the test. 7

CHOW TEST AND DUMMY VARIABLE GROUP TEST . reg COST N OCC NOCC Source | SS df MS Number of obs = 74 ---------+------------------------------ F( 3, 70) = 49.64 Model | 1.0009e+12 3 3.3363e+11 Prob > F = 0.0000 Residual | 4.7045e+11 70 6.7207e+09 R-squared = 0.6803 ---------+------------------------------ Adj R-squared = 0.6666 Total | 1.4713e+12 73 2.0155e+10 Root MSE = 81980 ------------------------------------------------------------------------------ COST | Coef. Std. Err. t P>|t| [95% Conf. Interval] ---------+-------------------------------------------------------------------- N | 152.2982 60.01932 2.537 0.013 32.59349 272.003 OCC | -3501.177 41085.46 -0.085 0.932 -85443.55 78441.19 NOCC | 284.4786 75.63211 3.761 0.000 133.6351 435.3221 _cons | 51475.25 31314.84 1.644 0.105 -10980.24 113930.7 We add an intercept dummy and a slope dummy to allow the overhead and marginal costs of the occupational schools to be different. Again we make a note of RSS, which is smaller than before. 8

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST occupational school regular school N Here are the regression lines for the two subsamples. 9

CHOW TEST AND DUMMY VARIABLE GROUP TEST Whole sample COST = 24,000 + 339N RSS = 8.91x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ To see if the cost functions are significantly different, we investigate whether there is a significant reduction in RSS when the dummy variables are added. 10

CHOW TEST AND DUMMY VARIABLE GROUP TEST Whole sample COST = 24,000 + 339N RSS = 8.91x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ We perform the F test described in the sequence on slope dummy variables. The numerator of the test statistic is the reduction in RSS on adding the dummy variables, divided by the cost in terms of degrees of freedom. 11

CHOW TEST AND DUMMY VARIABLE GROUP TEST Whole sample COST = 24,000 + 339N RSS = 8.91x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ The denominator is the RSS remaining after adding the dummy variables, divided by the number of degrees of freedom remaining. 12

CHOW TEST AND DUMMY VARIABLE GROUP TEST Whole sample COST = 24,000 + 339N RSS = 8.91x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ The critical value of F at the 0.1% level with 2 and 70 degrees of freedom is 7.6. Hence we conclude that the dummy variables do have significant explanatory power and the cost functions are different. 13

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST N With the Chow test approach we also start by running a regression using the whole sample, and make a note of the RSS. 14

CHOW TEST AND DUMMY VARIABLE GROUP TEST . reg COST N if OCC==0 Source | SS df MS Number of obs = 40 ---------+------------------------------ F( 1, 38) = 13.53 Model | 4.3273e+10 1 4.3273e+10 Prob > F = 0.0007 Residual | 1.2150e+11 38 3.1973e+09 R-squared = 0.2626 ---------+------------------------------ Adj R-squared = 0.2432 Total | 1.6477e+11 39 4.2249e+09 Root MSE = 56545 ------------------------------------------------------------------------------ COST | Coef. Std. Err. t P>|t| [95% Conf. Interval] ---------+-------------------------------------------------------------------- N | 152.2982 41.39782 3.679 0.001 68.49275 236.1037 _cons | 51475.25 21599.14 2.383 0.022 7750.064 95200.43 We then split the sample into occupational and regular schools, and run separate regressions, again making a note of RSS. This is the regression output when COST is regressed on N for the subsample of 40 regular schools. 15

CHOW TEST AND DUMMY VARIABLE GROUP TEST . reg COST N if OCC==1 Source | SS df MS Number of obs = 34 ---------+------------------------------ F( 1, 32) = 55.52 Model | 6.0538e+11 1 6.0538e+11 Prob > F = 0.0000 Residual | 3.4895e+11 32 1.0905e+10 R-squared = 0.6344 ---------+------------------------------ Adj R-squared = 0.6229 Total | 9.5433e+11 33 2.8919e+10 Root MSE = 1.0e+05 ------------------------------------------------------------------------------ COST | Coef. Std. Err. t P>|t| [95% Conf. Interval] ---------+-------------------------------------------------------------------- N | 436.7769 58.62085 7.451 0.000 317.3701 556.1836 _cons | 47974.07 33879.03 1.416 0.166 -21035.26 116983.4 And this is the regression output when COST is regressed on N using the subsample of 34 occupational schools. 16

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST occupational school regular school N The graph shows the regression lines. 17

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ The regression equations are as shown. 18

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ The cost functions are identical to those implicit in the dummy variable regression with both intercept and slope dummies. This is because the dummy variable regression has a dummy variable for each component of the original model (here, the constant and N). 19

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ Implicit cost function for regular schools COST = 51,000 + 152N ^ The intercept and the coefficient of N in the dummy variable regression are chosen so as to minimize the residual sum of squares for the reference category, the regular schools. Hence they must be the same as for the regression with regular schools only. 20

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ Implicit cost function for regular schools COST = 51,000 + 152N ^ Implicit cost function for occupational schools COST = 47,000 + 436N ^ The intercept and slope dummies then allow the intercept and slope coefficient to be modified so as to give the best possible fit for the occupational schools. Hence the implicit cost function must be the same as for the regression with occupational schools only. 21

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST occupational school regular school N The cost function for regular schools implicit in the dummy variable regression must coincide with the regression line for the regular schools only. 22

CHOW TEST AND DUMMY VARIABLE GROUP TEST COST occupational school regular school N Similarly, the cost function for occupational schools implicit in the dummy variable regression must coincide with the regression line for the occupational schools only. 23

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ Implicit cost function for regular schools COST = 51,000 + 152N ^ Implicit cost function for occupational schools COST = 47,000 + 436N ^ Since the cost functions implicit in the dummy variable regression coincide with those in the separate regressions, the residuals will be the same. It follows that RSS for the dummy variable regression must be equal the sum of RSS for the separate regressions. 24

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ Hence the F statistics for the F tests will be the same. The starting point for both approaches is the residual sum of squares for the basic regression making no distinction between types of school. 25

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ In the Chow test approach, RSS is reduced by splitting the sample. In the dummy variable approach, RSS is reduced by adding the intercept and slope dummies. RSS after making the change will be the same because the residuals will be the same. 26

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ This also means that the first part of the denominator of the F statistic will be the same. 27

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ ` The cost of the improvement in the fit is the same, since either way two extra parameters have to be estimated. 28

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ ` And either way, the number of degrees of freedom remaining will be 70, since the number of observations is 74 and 4 parameters have to be estimated. 29

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ Thus all the components of the F statistics are the same, and the outcome of the test must be the same. In this case, the null hypothesis of identical cost functions for the two types of school was rejected at the 0.1% level. 30

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ What are the advantages and disadvantages of the two approaches? 31

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ The Chow test is quick. You just run the three regressions and compute the test statistic. But it does not tell you how the functions differ, if they do. 32

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ The dummy variable approach involves more preparation because you have to define a dummy variable for the intercept and for each slope coefficient. 33

CHOW TEST AND DUMMY VARIABLE GROUP TEST Regular schools, subsample regression COST = 51,000 + 152N RSS = 1.22x1011 ^ Occupational schools, subsample regression COST = 47,000 + 436N RSS = 3.49x1011 ^ Whole sample, with dummy variables COST = 51,000 – 4,000OCC + 152N + 284NOCC RSS = 4.71x1011 ^ However, it is more informative because you can perform t tests on the individual dummy coefficients and find out where the functions differ, if they do. 34

CHOW TEST AND DUMMY VARIABLE GROUP TEST Basic model Y = b1 + b2X2 + b3X3+ … + bKXK + u Model with dummy variables Y = b1 + b2X2 + b3X3+ … + bKXK + dD + l2DX2 + l3DX3 + … + lKDXK + u D = 0 Y = b1 + b2X2 + b3X3+ … + bKXK + u D = 1 Y = (b1+d) + (b2+l2)X2 + (b3+l3)X3 + … + (bK+lK)XK + u A final note. The Chow test and the dummy variable group test are equivalent only if there is a full set of dummy variables. 35

CHOW TEST AND DUMMY VARIABLE GROUP TEST Basic model Y = b1 + b2X2 + b3X3+ … + bKXK + u Model with dummy variables Y = b1 + b2X2 + b3X3+ … + bKXK + dD + l2DX2 + l3DX3 + … + lKDXK + u D = 0 Y = b1 + b2X2 + b3X3+ … + bKXK + u D = 1 Y = (b1+d) + (b2+l2)X2 + (b3+l3)X3 + … + (bK+lK)XK + u By this is meant an intercept dummy (here D) and a slope dummy variable for every X (here DX2, DX3, … DXK). 36

CHOW TEST AND DUMMY VARIABLE GROUP TEST Basic model Y = b1 + b2X2 + b3X3+ … + bKXK + u Model with dummy variables Y = b1 + b2X2 + b3X3+ … + bKXK + dD + l2DX2 + l3DX3 + … + lKDXK + u D = 0 Y = b1 + b2X2 + b3X3+ … + bKXK + u D = 1 Y = (b1+d) + (b2+l2)X2 + (b3+l3)X3 + … + (bK+lK)XK + u If there is a full set of dummy variables, OLS will choose the intercept b1 and the b coefficients of X2 … XK so as to optimise the fit for the D = 0 observations. The coefficients will be exactly the same as if the regression has been run with only the subsample of D = 0 observations. 37

CHOW TEST AND DUMMY VARIABLE GROUP TEST Basic model Y = b1 + b2X2 + b3X3+ … + bKXK + u Model with dummy variables Y = b1 + b2X2 + b3X3+ … + bKXK + dD + l2DX2 + l3DX3 + … + lKDXK + u D = 0 Y = b1 + b2X2 + b3X3+ … + bKXK + u D = 1 Y = (b1+d) + (b2+l2)X2 + (b3+l3)X3 + … + (bK+lK)XK + u The coefficient of the intercept dummy D and the slope dummy variables will then be chosen so as to optimise the fit for the D = 1 observations. (b1+d), (b2+l2), …, (bK+lK) will be the same as the coefficients in a regression using only the subsample of D = 1 observations. 38

CHOW TEST AND DUMMY VARIABLE GROUP TEST Basic model Y = b1 + b2X2 + b3X3+ … + bKXK + u Model with dummy variables Y = b1 + b2X2 + b3X3+ … + bKXK + dD + l2DX2 + l3DX3 + … + lKDXK + u D = 0 Y = b1 + b2X2 + b3X3+ … + bKXK + u D = 1 Y = (b1+d) + (b2+l2)X2 + (b3+l3)X3 + … + (bK+lK)XK + u Thus with a full set of intercept and slope dummy variables, the improvement in fit on adding the dummy variables to the basic equation is the same as that obtained by splitting the sample and running separate subsample regressions. 39

CHOW TEST AND DUMMY VARIABLE GROUP TEST Basic model Y = b1 + b2X2 + b3X3+ … + bKXK + u Model with dummy variables Y = b1 + b2X2 + b3X3+ … + bKXK + dD + l2DX2 + l3DX3 + … + lKDXK + u D = 0 Y = b1 + b2X2 + b3X3+ … + bKXK + u D = 1 Y = (b1+d) + (b2+l2)X2 + (b3+l3)X3 + … + (bK+lK)XK + u It follows that the F statistic for the test of the joint explanatory power of the intercept and slope dummy variables is equivalent to the F statistic for the Chow test. 40

Copyright Christopher Dougherty 2012. These slideshows may be downloaded by anyone, anywhere for personal use. Subject to respect for copyright and, where appropriate, attribution, they may be used as a resource for teaching an econometrics course. There is no need to refer to the author. The content of this slideshow comes from Section 5.4 of C. Dougherty, Introduction to Econometrics, fourth edition 2011, Oxford University Press. Additional (free) resources for both students and instructors may be downloaded from the OUP Online Resource Centre http://www.oup.com/uk/orc/bin/9780199567089/. Individuals studying econometrics on their own who feel that they might benefit from participation in a formal course should consider the London School of Economics summer school course EC212 Introduction to Econometrics http://www2.lse.ac.uk/study/summerSchools/summerSchool/Home.aspx or the University of London International Programmes distance learning course EC2020 Elements of Econometrics www.londoninternational.ac.uk/lse. 2012.11.06