Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Managerial Economics in a Global Economy
Lesson 10: Linear Regression and Correlation
Inference for Regression
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Regression Analysis Simple Regression. y = mx + b y = a + bx.
1 Simple Linear Regression and Correlation The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES Assessing the model –T-tests –R-square.
Simple Regression. Major Questions Given an economic model involving a relationship between two economic variables, how do we go about specifying the.
Simple Linear Regression and Correlation
REGRESI SEDERHANA BAGIAN 2
Introduction to Regression Analysis
Estimating Demand Outline Where do demand functions come from? Sources of information for demand estimation Cross-sectional versus time series data Estimating.
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
Simple Linear Regression
Linear Regression and Correlation
The Simple Regression Model
SIMPLE LINEAR REGRESSION
Chapter Topics Types of Regression Models
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
Chapter 11 Multiple Regression.
SIMPLE LINEAR REGRESSION
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Lecture 19 Simple linear regression (Review, 18.5, 18.8)
Simple Linear Regression and Correlation
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression. Introduction In Chapters 17 to 19, we examine the relationship between interval variables via a mathematical equation. The motivation.
Correlation & Regression
Correlation and Linear Regression
SIMPLE LINEAR REGRESSION
Introduction to Linear Regression and Correlation Analysis
Linear Regression and Correlation
Correlation and Linear Regression
Regression Method.
5.1 Basic Estimation Techniques  The relationships we theoretically develop in the text can be estimated statistically using regression analysis,  Regression.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Correlation and Regression
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
© The McGraw-Hill Companies, Inc., Chapter 11 Correlation and Regression.
Production Planning and Control. A correlation is a relationship between two variables. The data can be represented by the ordered pairs (x, y) where.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Copyright © 2005 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics Thomas Maurice eighth edition Chapter 4.
10B11PD311 Economics REGRESSION ANALYSIS. 10B11PD311 Economics Regression Techniques and Demand Estimation Some important questions before a firm are.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
VI. Regression Analysis A. Simple Linear Regression 1. Scatter Plots Regression analysis is best taught via an example. Pencil lead is a ceramic material.
Lecture 10: Correlation and Regression Model.
Correlation & Regression Analysis
Linear Regression and Correlation Chapter GOALS 1. Understand and interpret the terms dependent and independent variable. 2. Calculate and interpret.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Chapter 13 Linear Regression and Correlation. Our Objectives  Draw a scatter diagram.  Understand and interpret the terms dependent and independent.
Inference about the slope parameter and correlation
The simple linear regression model and parameter estimation
Regression and Correlation
Correlation and Simple Linear Regression
Basic Estimation Techniques
Correlation and Simple Linear Regression
Basic Estimation Techniques
Correlation and Regression
Estimating Demand Outline Where do demand functions come from?
Correlation and Simple Linear Regression
Correlation and Regression
SIMPLE LINEAR REGRESSION
Simple Linear Regression and Correlation
SIMPLE LINEAR REGRESSION
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Y X 0 X and Y are not perfectly correlated. However, there is on average a positive relationship between Y and X X1X1 X2X2

11 Y1Y1 E(Y 1 /X 1 ) Y X0X1X1 E(Y i /X i ) =  0 +  1 X i We assume that expected conditional values of Y associated with alternative values of X fall on a line.  1 = Y 1 - E(Y 1 /X 1 )

âSpecification âEstimation âEvaluation âForecasting

Econometric models posit causal relationships among economic variables. Simple regression analysis is used to test the hypothesis about the relationship between a dependent variable (Y, or in our case, C) and independent variable (X, or in our case, Y)).  Our model is specified as follows: C = f (Y), where C: personal consumption expenditure Y: Personal disposable income

 Simple linear regression begins by plotting C-Y values (see table 1)on a scatter diagram (see figure 1) to determine if there exists an approximate linear relationship: (1) Since the data points are unlikely to fall exactly on a line, (1) must be modified to include a disturbance term (u i ) (2)  0 and  1 are called parameters or population parameters.  We estimate these parameters using the data we have available

Table 1

 We estimate the values of  0 and  1 using the Ordinary Least Squares (OLS) method. OLS is a technique for fitting the "best" straight line to the sample of XY observations. The line of best fit is that which minimizes the sum of the squared (vertical) deviations of the sample points from the line: Where, C i are the actual observations of consumption are fitted values of consumption

C Y C1C1 Y1Y1 0 e1e1

The OLS estimators--single variable case are estimators of the true parameters  0 and  1 Note that we use X to denote the explanatory variable and Y is the dependent variable.

Table 2

Thus, we have: Thus the equation obtained from the regression is:

Table 3: Fitted values of consumption

 Goodness of fit criteria Standard errors of the estimates Are the estimates statistically significant? Constructing confidence intervals The coefficient of determination (R 2 ). The standard error of the regression These statistics tell us how well the equation obtained from the regression performs in terms of producing accurate forecasts

We assume that the regression coefficients are normally distributed variables. The standard error (or standard deviation) of the estimates is a measure of the dispersion of the estimates around their mean value. As a general principle, the smaller the standard error, the better the estimates (in terms of yielding accurate forecasts of the dependent variable). The following rule-of-thumb is useful:"[the] standard error of the regression coefficient should be less than half of the size of [the] corresponding regression coefficient." Let denote the standard error of our estimate of the slope parameter

By reference to the SPSS output, we see that the standard error of our estimate of  1 is 0.049, whereas our estimate of  1 is Hence our estimate is about 17 times the size of its standard error Note that:

ß To test for the significance of our estimate of  1, we set the following null hypothesis, H 0, and the alternative hypothesis, H 1 ßH 0 :  1  0 ßH 1 :  1 > 0 ßThe t distribution is used to test for statistical significance of the estimate: The t test is a way of comparing the error suggested by the null hypothesis to the standard error of the estimate A rule-of thumb: if t > 2, reject H 0

Constructing confidence intervals To find the 95 percent confidence interval for  1, that is: Pr( a <  1 < b) =.95 To find the upper and lower boundaries of the confidence interval (a and b): Where t c is the critical value of t at the 5 percent confidence level (two-sided,10 degrees of freedom ). t c = Working it out, we have: Pr(.752<  1 <.970) =.95  We can be 95 percent confident that the true value of the slope coefficient is in this range.

The coefficient of determination (R 2 )  The coefficient of determination, R 2, is defined as the proportion of the total variation in the dependent variable (Y) "explained" by the regression of Y on the independent variable (X). The total variation in Y or the total sum of squares (TSS) is defined as:  The explained variation in the dependent variable(Y) is called the regression sum of squares (RSS) and is given by: Note:

What remains is the unexplained variation in the dependent variable or the error sum of squares (ESS) We can say the following: TSS = RSS + ESS, or Total variation = Explained variation + Unexplained variation R 2 is defined as:

 Note that: 0  R 2  1  If R 2 = 0, all the sample points lie on a horizontal line or in a circle  If R 2 = 1, the sample points all lie on the regression line  In our case, R 2  0.984, meaning that 98.4 percent of the variation in the dependent variable (consumption) is explained by the regression. Think of R 2 as the proportion of the total deviation of the dependent variable from its mean value that is accounted for by the explanatory variable(s).

 The standard error of the regression (s) is given by  In our case, s = 3.40  Regression is based on the assumption that the error term is normally distributed, so that 6.87% of the actual values of the dependent variable should be within one standard error (  $3.4 billion in our example) of their fitted value.  Also, 95.45% of the observed values of consumption should be within 2 standard errors of their fitted values (  $6.8 billion).

Our forecasting equation was estimated as follows:  At the most basic level, forecasting consists of inserting forecasted values of the explanatory variable X (disposable income) into the forecasting equation to obtain forecasted values of the dependent variable Y (personal consumption expenditure).

Our ability to generate accurate forecasts of the dependent variable depends on two factors:  Do we have good forecasts of the explanatory variable?  Does our model exhibit structural stability, i.e., will the causal relationship between C and Y expressed in our forecasting equation hold up over time? After all, the estimated coefficients are average values for a specific time interval ( ). While the past may be a serviceable guide to the future in the case of purely physical phenomena, the same principle does not necessarily hold in the realm of social phenomena (to which economy belongs). Can we make a good forecast?

Having forecasted values of income in hand, we can forecast consumption through the year 2003