Part 8: Hypothesis Testing 8-1/50 Econometrics I Professor William Greene Stern School of Business Department of Economics.

Slides:



Advertisements
Similar presentations
The Multiple Regression Model.
Advertisements

Part 12: Asymptotics for the Regression Model 12-1/39 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Hypothesis Testing Steps in Hypothesis Testing:
Econ 488 Lecture 5 – Hypothesis Testing Cameron Kaplan.
3. Binary Choice – Inference. Hypothesis Testing in Binary Choice Models.
Part 7: Estimating the Variance of b 7-1/53 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Lecture 8 Relationships between Scale variables: Regression Analysis
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Chapter 13 Multiple Regression
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
Chapter 10 Simple Regression.
Chapter 12 Multiple Regression
The Simple Regression Model
Stat 112 – Notes 3 Homework 1 is due at the beginning of class next Thursday.
Economics 20 - Prof. Anderson1 Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
T-test.
Chapter 11 Multiple Regression.
Horng-Chyi HorngStatistics II41 Inference on the Mean of a Population - Variance Known H 0 :  =  0 H 0 :  =  0 H 1 :    0, where  0 is a specified.
Part 18: Regression Modeling 18-1/44 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics.
Inferences About Process Quality
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Part 7: Multiple Regression Analysis 7-1/54 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
Introduction to Regression Analysis, Chapter 13,
Part 24: Multiple Regression – Part /45 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department.
AM Recitation 2/10/11.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
CENTRE FOR INNOVATION, RESEARCH AND COMPETENCE IN THE LEARNING ECONOMY Session 2: Basic techniques for innovation data analysis. Part I: Statistical inferences.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Interval Estimation and Hypothesis Testing
Microeconometric Modeling William Greene Stern School of Business New York University.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Part 9: Hypothesis Testing /29 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Part 2: Model and Inference 2-1/49 Regression Models Professor William Greene Stern School of Business IOMS Department Department of Economics.
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005 Dr. John Lipp Copyright © Dr. John Lipp.
Interval Estimation and Hypothesis Testing Prepared by Vera Tabakova, East Carolina University.
Correlation Assume you have two measurements, x and y, on a set of objects, and would like to know if x and y are related. If they are directly related,
Statistical Inference for the Mean Objectives: (Chapter 9, DeCoursey) -To understand the terms: Null Hypothesis, Rejection Region, and Type I and II errors.
Lecture 11 Preview: Hypothesis Testing and the Wald Test Wald Test Let Statistical Software Do the Work Testing the Significance of the “Entire” Model.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
1Prof. Dr. Rainer Stachuletz Multiple Regression Analysis y =  0 +  1 x 1 +  2 x  k x k + u 2. Inference.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University.
Created by Erin Hodgess, Houston, Texas Section 7-1 & 7-2 Overview and Basics of Hypothesis Testing.
Statistical Inference for the Mean Objectives: (Chapter 8&9, DeCoursey) -To understand the terms variance and standard error of a sample mean, Null Hypothesis,
Class Seven Turn In: Chapter 18: 32, 34, 36 Chapter 19: 26, 34, 44 Quiz 3 For Class Eight: Chapter 20: 18, 20, 24 Chapter 22: 34, 36 Read Chapters 23 &
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Inference about the slope parameter and correlation
Chapter 5 STATISTICAL INFERENCE: ESTIMATION AND HYPOTHESES TESTING
Lecture 8 Preview: Interval Estimates and Hypothesis Testing
Statistics and Data Analysis
6-1 Introduction To Empirical Models
Review for Exam 2 Some important themes from Chapters 6-9
Discrete Event Simulation - 4
Review of Statistical Inference
Interval Estimation and Hypothesis Testing
Chengyuan Yin School of Mathematics
Tutorial 1: Misspecification
Chapter 7: The Normality Assumption and Inference with OLS
Chengyuan Yin School of Mathematics
Inference on the Mean of a Population -Variance Known
Econometrics I Professor William Greene Stern School of Business
Microeconometric Modeling
Econometrics I Professor William Greene Stern School of Business
Econometrics I Professor William Greene Stern School of Business
Presentation transcript:

Part 8: Hypothesis Testing 8-1/50 Econometrics I Professor William Greene Stern School of Business Department of Economics

Part 8: Hypothesis Testing 8-2/50 Econometrics I Part 8 – Interval Estimation and Hypothesis Testing

Part 8: Hypothesis Testing 8-3/50 Interval Estimation  b = point estimator of   We acknowledge the sampling variability. Estimated sampling variance b =  + sampling variability induced by 

Part 8: Hypothesis Testing 8-4/50  Point estimate is only the best single guess  Form an interval, or range of plausible values  Plausible  likely values with acceptable degree of probability.  To assign probabilities, we require a distribution for the variation of the estimator.  The role of the normality assumption for 

Part 8: Hypothesis Testing 8-5/50 Confidence Interval b k = the point estimate Std.Err[b k ] = sqr{[σ 2 (XX) -1 ] kk } = v k Assume normality of ε for now: b k ~ N[β k,v k 2 ] for the true β k. (b k -β k )/v k ~ N[0,1] Consider a range of plausible values of β k given the point estimate b k. b k  sampling error. Measured in standard error units, |(b k – β k )/ v k | < z* Larger z*  greater probability (“confidence”) Given normality, e.g., z* = 1.96  95%, z*=1.645  90% Plausible range for β k then is b k ± z* v k

Part 8: Hypothesis Testing 8-6/50 Computing the Confidence Interval Assume normality of ε for now: b k ~ N[β k,v k 2 ] for the true β k. (b k -β k )/v k ~ N[0,1] v k = [σ 2 (X’X) -1 ] kk is not known because σ 2 must be estimated. Using s 2 instead of σ 2, (b k -β k )/Est.(v k ) ~ t[N-K]. (Proof: ratio of normal to sqr(chi-squared)/df is pursued in your text.) Use critical values from t distribution instead of standard normal. Will be the same as normal if N > 100.

Part 8: Hypothesis Testing 8-7/50 Confidence Interval Critical t[.975,29] = Confidence interval based on t:  *.1501 Confidence interval based on normal:  *.1501

Part 8: Hypothesis Testing 8-8/50 Bootstrap Confidence Interval For a Coefficient

Part 8: Hypothesis Testing 8-9/50 Bootstrap CI for Least Squares

Part 8: Hypothesis Testing 8-10/50 Bootstrap CI for Least Absolute Deviations

Part 8: Hypothesis Testing 8-11/50 Testing a Hypothesis Using a Confidence Interval Given the range of plausible values Testing the hypothesis that a coefficient equals zero or some other particular value: Is the hypothesized value in the confidence interval? Is the hypothesized value within the range of plausible values? If not, reject the hypothesis.

Part 8: Hypothesis Testing 8-12/50 Test a Hypothesis About a Coefficient

Part 8: Hypothesis Testing 8-13/50 Classical Hypothesis Testing We are interested in using the linear regression to support or cast doubt on the validity of a theory about the real world counterpart to our statistical model. The model is used to test hypotheses about the underlying data generating process.

Part 8: Hypothesis Testing 8-14/50 Types of Tests  Nested Models: Restriction on the parameters of a particular model y =  1 +  2 x +  3 z + ,  3 = 0  Nonnested models: E.g., different RHS variables y t =  1 +  2 x t +  3 x t-1 +  t y t =  1 +  2 x t +  3 y t-1 + w t  Specification tests:  ~ N[0,  2 ] vs. some other distribution

Part 8: Hypothesis Testing 8-15/50 Methodology  Bayesian Prior odds compares strength of prior beliefs in two states of the world Posterior odds compares revised beliefs Symmetrical treatment of competing ideas Not generally practical to carry out in meaningful situations  Classical “Null” hypothesis given prominence Propose to “reject” toward default favor of “alternative” Asymmetric treatment of null and alternative Huge gain in practical applicability

Part 8: Hypothesis Testing 8-16/50 Neyman – Pearson Methodology  Formulate null and alternative hypotheses  Define “Rejection” region = sample evidence that will lead to rejection of the null hypothesis.  Gather evidence  Assess whether evidence falls in rejection region or not.

Part 8: Hypothesis Testing 8-17/50 Inference in the Linear Model Formulating hypotheses: linear restrictions as a general framework Hypothesis Testing J linear restrictions Analytical framework:y = X  +  Hypothesis:R  - q = 0, Substantive restrictions: What is a "testable hypothesis?"  Substantive restriction on parameters  Reduces dimension of parameter space  Imposition of restriction degrades estimation criterion

Part 8: Hypothesis Testing 8-18/50 Testable Implications of a Theory  Investors care about interest rates and expected inflation: I = b 1 + b 2 r + b 3 dp + e  Investors care about real interest rates I = c 1 + c 2 (r-dp) + c 3 dp + e No testable restrictions implied. c 1 =b 1, c 2 =b 2 -b 3, c 3 =b 3.  Investors care only about real interest rates I = f 1 + f 2 (r-dp) + f 3 dp + e. f 3 = 0

Part 8: Hypothesis Testing 8-19/50 The General Linear Hypothesis: H 0 : R  - q = 0 A unifying departure point: Regardless of the hypothesis, least squares is unbiased. E[b] =  The hypothesis makes a claim about the population R  – q = 0. Then, if the hypothesis is true, E[Rb – q] = 0. The sample statistic, Rb – q will not equal zero. Two possibilities: Rb – q is small enough to attribute to sampling variability Rb – q is too large (by some measure) to be plausibly attributed to sampling variability Large Rb – q is the rejection region.

Part 8: Hypothesis Testing 8-20/50 Approaches to Defining the Rejection Region (1) Imposing the restrictions leads to a loss of fit. R 2 must go down. Does it go down “a lot?” (I.e., significantly?). R u 2 = unrestricted model, R r 2 = restricted model fit. F = { (R u 2 – R r 2 )/J } / [(1 – R u 2 )/(N-K)] = F[J,N-K]. (2) Is Rb - q close to 0? Basing the test on the discrepancy vector: m = Rb - q. Using the Wald criterion: m(Var[m]) -1 m has a chi-squared distribution with J degrees of freedom But, Var[m] = R[  2 (X’X) -1 ]R. If we use our estimate of  2, we get an F[J,N-K], instead. (Note, this is based on using ee/(N-K) to estimate  2.) These are the same for the linear model

Part 8: Hypothesis Testing 8-21/50 Testing Fundamentals - I  SIZE of a test = Probability it will incorrectly reject a “true” null hypothesis.  This is the probability of a Type I error. Under the null hypothesis, F(3,100) has an F distribution with (3,100) degrees of freedom. Even if the null is true, F will be larger than the 5% critical value of 2.7 about 5% of the time.

Part 8: Hypothesis Testing 8-22/50 Testing Procedures How to determine if the statistic is 'large.' Need a 'null distribution.' If the hypothesis is true, then the statistic will have a certain distribution. This tells you how likely certain values are, and in particular, if the hypothesis is true, then 'large values' will be unlikely. If the observed statistic is too large, conclude that the assumed distribution must be incorrect and the hypothesis should be rejected. For the linear regression model with normally distributed disturbances, the distribution of the relevant statistic is F with J and N-K degrees of freedom.

Part 8: Hypothesis Testing 8-23/50 Distribution Under the Null

Part 8: Hypothesis Testing 8-24/50 A Simulation Experiment sample ; $ matrix ; fvalues=init(1000,1,0)$ proc$ create ; fakelogc = rnn( , )$ Coefficients all = 0 regress ; quietly ; lhs = fakelogc Compute regression ; rhs=one,logq,logpl_pf,logpk_pf$ calc ; fstat = (rsqrd/3)/((1-rsqrd)/(n-4))$ Compute F matrix ; fvalues(i)=fstat$ Save 1000 Fs endproc execute ; i= 1,1000 $ 1000 replications histogram ; rhs = fvalues ; title=F Statistic for H0:b2=b3=b4=0$

Part 8: Hypothesis Testing 8-25/50 Simulation Results 48 outcomes to the right of 2.7 in this run of the experiment.

Part 8: Hypothesis Testing 8-26/50 Testing Fundamentals - II  POWER of a test = the probability that it will correctly reject a “false null” hypothesis  This is 1 – the probability of a Type II error.  The power of a test depends on the specific alternative.

Part 8: Hypothesis Testing 8-27/50 Power of a Test Null: Mean = 0. Reject if observed mean Prob(Reject null|mean=0) = 0.05 Prob(Reject null|mean=.5)= Prob(Reject null|mean=1)= Increases as the (alternative) mean rises.

Part 8: Hypothesis Testing 8-28/50 Test Statistic For the fit measures, use a normalized measure of the loss of fit:

Part 8: Hypothesis Testing 8-29/50 Test Statistics Forming test statistics: For distance measures use Wald type of distance measure, W = m[Est.Var(m)] -1 m An important relationship between t and F For a single restriction, m = r’b - q. The variance is r’(Var[b])r The distance measure is m / standard error of m.

Part 8: Hypothesis Testing 8-30/50 An important relationship between t and F For a single restriction, F[1,N-K] is the square of the t ratio.

Part 8: Hypothesis Testing 8-31/50 Application Time series regression, LogG =  1 +  2 logY +  3 logPG +  4 logPNC +  5 logPUC +  6 logPPT +  7 logPN +  8 logPD +  9 logPS +  Period = Note that all coefficients in the model are elasticities.

Part 8: Hypothesis Testing 8-32/50 Full Model Ordinary least squares regression LHS=LG Mean = Standard deviation = Number of observs. = 36 Model size Parameters = 9 Degrees of freedom = 27 Residuals Sum of squares = <******* Standard error of e = <******* Fit R-squared = <******* Adjusted R-squared = <******* Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| *** LY| *** LPG| *** LPNC| LPUC| * LPPT| LPN| *** LPD|.92018*** LPS| ***

Part 8: Hypothesis Testing 8-33/50 Test About One Parameter Is the price of public transportation really relevant? H 0 :  6 = 0. Confidence interval: b 6  t(.95,27)  Standard error =  2.052(.07859) =  = ( ,.27698) Contains 0.0. Do not reject hypothesis Regression fit if drop? Without LPPT, R-squared= Compare R 2, was.99605, F(1,27) = [( )/1]/[( )/(36-9)] = = (with some rounding difference)

Part 8: Hypothesis Testing 8-34/50 Hypothesis Test: Sum of Coefficients = 1? Ordinary least squares regression LHS=LG Mean = Standard deviation = Number of observs. = 36 Model size Parameters = 9 Degrees of freedom = 27 Residuals Sum of squares = <******* Standard error of e = <******* Fit R-squared = <******* Adjusted R-squared = <******* Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| *** LY| *** LPG| *** LPNC| LPUC| * LPPT| LPN| *** LPD|.92018*** LPS| ***

Part 8: Hypothesis Testing 8-35/50 Imposing the Restriction Linearly restricted regression LHS=LG Mean = Standard deviation = Number of observs. = 36 Model size Parameters = 8 <*** 9 – 1 restriction Degrees of freedom = 28 Residuals Sum of squares = <*** With the restriction Residuals Sum of squares = <*** Without the restriction Fit R-squared = Restrictns. F[ 1, 27] (prob) = 8.5(.01) Not using OLS or no constant.R2 & F may be < Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| *** LY| *** LPG| *** LPNC|.46945*** LPUC| LPPT|.24223*** LPN| *** LPD| LPS| *** F = [( )/1] / [ /(36 – 9)] =

Part 8: Hypothesis Testing 8-36/50 Joint Hypotheses Joint hypothesis: Income elasticity = +1, Own price elasticity = -1. The hypothesis implies that logG = β 1 + logY – logPg + β 4 logPNC +... Strategy: Regress logG – logY + logPg on the other variables and Compare the sums of squares With two restrictions imposed Residuals Sum of squares = Fit R-squared = Unrestricted Residuals Sum of squares = Fit R-squared = F = (( )/2) / ( /(36-9)) = The critical F for 95% with 2,27 degrees of freedom is The hypothesis is rejected.

Part 8: Hypothesis Testing 8-37/50 Basing the Test on R 2 After building the restrictions into the model and computing restricted and unrestricted regressions: Based on R 2 s, F = (( )/2)/(( )/(36-9)) = (!) What's wrong? The unrestricted model used LHS = logG. The restricted one used logG-logY. The calculation is safe using the sums of squared residuals.

Part 8: Hypothesis Testing 8-38/50 Wald Distance Measure Testing more generally about a single parameter. Sample estimate is b k Hypothesized value is β k How far is β k from b k ? If too far, the hypothesis is inconsistent with the sample evidence. Measure distance in standard error units t = (b k - β k )/Estimated v k. If t is “large” (larger than critical value), reject the hypothesis.

Part 8: Hypothesis Testing 8-39/50 The Wald Statistic

Part 8: Hypothesis Testing 8-40/50 Robust Tests  The Wald test generally will (when properly constructed) be more robust to failures of the narrow model assumptions than the t or F  Reason: Based on “robust” variance estimators and asymptotic results that hold in a wide range of circumstances.  Analysis: Later in the course – after developing asymptotics.

Part 8: Hypothesis Testing 8-41/50 Particular Cases Some particular cases: One coefficient equals a particular value: F = [(b - value) / Standard error of b ] 2 = square of familiar t ratio. Relationship is F [ 1, d.f.] = t 2 [d.f.] A linear function of coefficients equals a particular value (linear function of coefficients - value) 2 F = Variance of linear function Note square of distance in numerator Suppose linear function is  k w k b k Variance is  k  l w k w l Cov[b k,b l ] This is the Wald statistic. Also the square of the somewhat familiar t statistic. Several linear functions. Use Wald or F. Loss of fit measures may be easier to compute.

Part 8: Hypothesis Testing 8-42/50 Hypothesis Test: Sum of Coefficients Do the three aggregate price elasticities sum to zero? H 0 :β 7 + β 8 + β 9 = 0 R = [0, 0, 0, 0, 0, 0, 1, 1, 1], q = [0] Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X LPN| *** LPD|.92018*** LPS| ***

Part 8: Hypothesis Testing 8-43/50 Wald Test

Part 8: Hypothesis Testing 8-44/50 Using the Wald Statistic --> Matrix ; R = [0,1,0,0,0,0,0,0,0 / 0,0,1,0,0,0,0,0,0]$ --> Matrix ; q = [1/-1]$ --> Matrix ; list ; m = R*b - q $ Matrix M has 2 rows and 1 columns | | > Matrix ; list ; vm = R*varb*R' $ Matrix VM has 2 rows and 2 columns | | > Matrix ; list ; w = m' m $ Matrix W has 1 rows and 1 columns | Joint hypothesis: b(LY) = 1 b(LPG) = -1

Part 8: Hypothesis Testing 8-45/50 Application: Cost Function

Part 8: Hypothesis Testing 8-46/50 Regression Results Ordinary least squares regression LHS=C Mean = Standard deviation = Number of observs. = 158 Model size Parameters = 9 Degrees of freedom = 149 Residuals Sum of squares = Standard error of e = Fit R-squared = Adjusted R-squared = Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| K| L| *** F| Q| *** Q2|.05250*** QK| QL|.11698*** QF|.05950** Note: ***, **, * = Significance at 1%, 5%, 10% level

Part 8: Hypothesis Testing 8-47/50 Price Homogeneity: Only Price Ratios Matter β 2 + β 3 + β 4 = 1. β 7 + β 8 + β 9 = Linearly restricted regression LHS=C Mean = Standard deviation = Number of observs. = 158 Model size Parameters = 7 Degrees of freedom = 151 Residuals Sum of squares = Standard error of e = Fit R-squared = Restrictns. F[ 2, 149] (prob) = 8.5(.0003) Not using OLS or no constant. Rsqrd & F may be < Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| *** K|.38943** L| F|.41927** Q|.45889*** Q2|.06008*** QK| QL| QF|

Part 8: Hypothesis Testing 8-48/50 Imposing the Restrictions Alternatively, compute the restricted regression by converting to price ratios and Imposing the restrictions directly. This is a regression of log(c/pf) on log(pk/pf), log(pl/pf) etc Ordinary least squares regression LHS=LOGC_PF Mean = Standard deviation = Number of observs. = 158 Model size Parameters = 7 Degrees of freedom = 151 Residuals Sum of squares = (restricted) Residuals Sum of squares = (unrestricted) Variable| Coefficient Standard Error t-ratio P[|T|>t] Mean of X Constant| *** LOGQ|.45889*** LOGQSQ|.06008*** LOGPK_PF|.38943** LOGPL_PF| LQ_LPKPF| LQ_LPLPF| F Statistic = (( – )/2) / ( /(158-9)) = 8.52 (This is a problem. The underlying theory requires linear homogeneity in prices.)

Part 8: Hypothesis Testing 8-49/50 Wald Test of the Restrictions Chi squared = J*F wald ; fn1 = b_k + b_l + b_f - 1 ; fn2 = b_qk + b_ql + b_qf – 0 $ WALD procedure. Estimates and standard errors for nonlinear functions and joint test of nonlinear restrictions. Wald Statistic = Prob. from Chi-squared[ 2] = Variable| Coefficient Standard Error b/St.Er. P[|Z|>z] Fncn(1)| *** Fncn(2)|.21921*** Note: ***, **, * = Significance at 1%, 5%, 10% level Recall: F Statistic = (( – )/2) / ( /(158-9)) = 8.52 The chi squared statistic is JF. Here, 2F = = the chi squared.

Part 8: Hypothesis Testing 8-50/50 Test of Homotheticity Cross Product Terms = 0 Homothetic Production: β 7 = β 8 = β 9 = 0. With linearity homogeneity in prices already imposed, this is β 6 = β 7 = WALD procedure. Estimates and standard errors for nonlinear functions and joint test of nonlinear restrictions. Wald Statistic = Prob. from Chi-squared[ 2] = Variable| Coefficient Standard Error b/St.Er. P[|Z|>z] Fncn(1)| Fncn(2)| The F test would produce F = (( – )/1)/( /(158-7)) = 1.285