Presentation is loading. Please wait.

Presentation is loading. Please wait.

FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u.

Similar presentations


Presentation on theme: "FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u."— Presentation transcript:

1 FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u

2 FIN357 Li2 Some Terminology In the simple linear regression model, where y =  0 +  1 x + u, we typically refer to y as the Dependent Variable, or Left-Hand Side Variable, or Explained Variable, or

3 FIN357 Li3 Some Terminology we typically refer to x as the Independent Variable, or Right-Hand Side Variable, or Explanatory Variable, or Regressor, or Control Variables

4 FIN357 Li4 Some Assumptions The average value of u, the error term, in the population is 0. That is, E(u) = 0 E(u|x) = 0 E(y|x) =  0 +  1 x

5 FIN357 Li5.. x1x1 x2x2 E(y|x) as a linear function of x, where for any x the distribution of y is centered about E(y|x) E(y|x) =  0 +  1 x y f(y)

6 FIN357 Li6 Ordinary Least Squares (OLS) Let {(x i,y i ): i=1, …,n} denote a random sample of size n from the population For each observation in this sample, it will be the case that y i =  0 +  1 x i + u i

7 FIN357 Li7.... y4y4 y1y1 y2y2 y3y3 x1x1 x2x2 x3x3 x4x4 } } { { u1u1 u2u2 u3u3 u4u4 x y Population regression line, sample data points and the associated error terms E(y|x) =  0 +  1 x

8 FIN357 Li8 Basic idea of regression is to estimate the population parameters from a sample Intuitively, OLS is fitting a line through the sample points such that the sum of squared residuals is as small as possible. The residual, û, is an estimate of the error term, u, and is the difference between the fitted line (sample regression function) and the sample point

9 FIN357 Li9.... y4y4 y1y1 y2y2 y3y3 x1x1 x2x2 x3x3 x4x4 } } { { û1û1 û2û2 û3û3 û4û4 x y Sample regression line, sample data points and the associated estimated error terms (residuals)

10 FIN357 Li10 It could be shown that estimated coefficient is

11 FIN357 Li11 OLS regressions Now that we’ve derived the formula for calculating the OLS estimates of our parameters (also called coefficients), you’ll be happy to know you don’t have to compute them by hand Regressions in GRETL are very simple. Have you installed the software yet?

12 FIN357 Li12 Some Properties of OLS If x and y are positively correlated, the slope will be positive If x and y are negatively correlated, the slope will be negative The sum of the OLS residuals is zero The OLS regression line always goes through the mean of the sample

13 FIN357 Li13 More terminology

14 FIN357 Li14 Goodness-of-Fit How do we think about how well our sample regression line fits our sample data? Can compute the fraction of the total sum of squares (SST) that is explained by the model, call this the R-squared of regression R 2 = SSE/SST = 1 – SSR/SST

15 FIN357 Li15 Under some conditions, OLS esimated coefficients are unbiased. Unbiasedness is a description of the estimator “on average” In a given sample we may be “near” or “far” from the true parameter


Download ppt "FIN357 Li1 The Simple Regression Model y =  0 +  1 x + u."

Similar presentations


Ads by Google