© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Hypothesis Testing Steps in Hypothesis Testing:
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Objectives (BPS chapter 24)
Chapter 12 Simple Linear Regression
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
© 2010 Pearson Prentice Hall. All rights reserved Regression Interval Estimates.
Hypothesis Testing Using a Single Sample
© 2010 Pearson Prentice Hall. All rights reserved Single Factor ANOVA.
© 2010 Pearson Prentice Hall. All rights reserved Two Sample Hypothesis Testing for Means from Independent Groups.
Two Sample Hypothesis Testing for Proportions
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
© 2010 Pearson Prentice Hall. All rights reserved Hypothesis Testing Using a Single Sample.
© 2010 Pearson Prentice Hall. All rights reserved Two Sample Hypothesis Testing for Means from Paired or Dependent Groups.
Chapter 10 Simple Regression.
Linear Regression and Correlation Analysis
Simple Linear Regression Analysis
SIMPLE LINEAR REGRESSION
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Simple Linear Regression and Correlation
Chapter 12 Section 1 Inference for Linear Regression.
Simple Linear Regression Analysis
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Lecture 5 Correlation and Regression
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
SIMPLE LINEAR REGRESSION
Correlation and Regression
Chapter 13: Inference in Regression
Chapter 11 Simple Regression
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Chapter Correlation and Regression 1 of 84 9 © 2012 Pearson Education, Inc. All rights reserved.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 2 – Slide 1 of 25 Chapter 11 Section 2 Inference about Two Means: Independent.
Inference for Linear Regression Conditions for Regression Inference: Suppose we have n observations on an explanatory variable x and a response variable.
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Chapter 9 Hypothesis Testing and Estimation for Two Population Parameters.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
Statistics for Business and Economics 7 th Edition Chapter 11 Simple Regression Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Introduction to Linear Regression
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Section Inference about Two Means: Independent Samples 11.3.
1 Lecture 4 Main Tasks Today 1. Review of Lecture 3 2. Accuracy of the LS estimators 3. Significance Tests of the Parameters 4. Confidence Interval 5.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
+ Chapter 12: More About Regression Section 12.1 Inference for Linear Regression.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
STA 286 week 131 Inference for the Regression Coefficient Recall, b 0 and b 1 are the estimates of the slope β 1 and intercept β 0 of population regression.
Applied Quantitative Analysis and Practices LECTURE#25 By Dr. Osman Sadiq Paracha.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 26 Chapter 11 Section 1 Inference about Two Means: Dependent Samples.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics Seventh Edition By Brase and Brase Prepared by: Lynn Smith.
Chapter 12 Simple Linear Regression n Simple Linear Regression Model n Least Squares Method n Coefficient of Determination n Model Assumptions n Testing.
© 2010 Pearson Prentice Hall. All rights reserved Chapter Hypothesis Tests Regarding a Parameter 10.
Lesson Testing the Significance of the Least Squares Regression Model.
© 2010 Pearson Prentice Hall. All rights reserved Chapter Hypothesis Tests Regarding a Parameter 10.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Hypothesis Tests Regarding a Parameter 10.
© 2010 Pearson Prentice Hall. All rights reserved Chapter Hypothesis Tests Regarding a Parameter 10.
Lecture #25 Tuesday, November 15, 2016 Textbook: 14.1 and 14.3
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Inference for Regression (Chapter 14) A.P. Stats Review Topic #3
Correlation and Regression
Chapter 12 Inference on the Least-squares Regression Line; ANOVA
St. Edward’s University
Presentation transcript:

© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models

14-2 The least-squares regression model is given by where y i is the value of the response variable for the i th individual  0 and  1 are the parameters to be estimated based on sample data x i is the value of the explanatory variable for the i th individual  i is a random error term with mean 0 and variance, the error terms are independent and normally distributed. i=1,…,n, where n is the sample size (number of ordered pairs in the data set)

Formulas for the slope and intercept estimates. For the estimated regression equation given by the formula: The slope b 1 is calculated by: And the intercept b 0 can be found with :

14-4 The standard error of the estimate, s e, is found using the formula

14-5 Parallel Example 2: Compute the Standard Error Compute the standard error of the estimate for the drilling data which is presented on the next slide.

14-6

14-7 Solution Step 1: Using technology (i.e. Minitab), we find the least squares regression line to be Step 2, 3: The predicted values as well as the residuals for the 12 observations are given in the table on the next slide

14-8

14-9 Solution Step 4: We find the sum of the squared residuals by summing the last column of the table: Step 5: The standard error of the estimate is then given by:

14-10 CAUTION! Be sure to divide by n-2 when computing the standard error of the estimate.

14-11 Parallel Example 4: Compute the Standard Error Verify that the residuals from the drilling example are normally distributed.

14-12

Conclusion: We have insufficient evidence at the 5% level of significance to support the claim that the residual errors from this model are not normally distributed.

14-14 Hypothesis Test Regarding the Slope Coefficient,  1 To test whether two quantitative variables are linearly related, we use the following steps provided that 1.the sample is obtained using random sampling. 2.the residuals are normally distributed with constant error variance.

14-15 Step 1: Determine the null and alternative hypotheses. The hypotheses can be structured in one of three ways: Step 2: Select a level of significance, , depending on the seriousness of making a Type I error. Two-tailedLeft-TailedRight-Tailed H 0 :  1 = 0 H 1 :  1  0H 1 :  1 < 0H 1 :  1 > 0

14-16 Step 3: Compute the test statistic which follows Student’s t-distribution with n-2 degrees of freedom. Remember, when computing the test statistic, we assume the null hypothesis to be true. So, we assume that  1 =0.

14-17 Step 4: Use Table VI to estimate the P-value using n-2 degrees of freedom. P-Value Approach

14-18 P-Value Approach Two-Tailed

14-19 P-Value Approach Left-Tailed

14-20 P-Value Approach Right-Tailed

14-21 Step 5: If the P-value < , reject the null hypothesis. P-Value Approach

14-22 Step 6: State the conclusion.

14-23 CAUTION! Before testing H 0 :  1 = 0, be sure to draw a residual plot to verify that a linear model is appropriate.

14-24 Parallel Example 5: Testing for a Linear Relation Test the claim that there is a linear relation between drill depth and drill time at the  = 0.05 level of significance using the drilling data.

14-25 Solution Verify the requirements: We assume that the experiment was randomized so that the data can be assumed to represent a random sample. In Parallel Example 4 we confirmed that the residuals were normally distributed by constructing a normal probability plot. To verify the requirement of constant error variance, we plot the residuals against the explanatory variable, drill depth.

14-26 There is no discernable pattern.

14-27 Solution Step 1: We want to determine whether a linear relation exists between drill depth and drill time without regard to the sign of the slope. This is a two-tailed test with H 0 :  1 = 0 versus H 1 :  1  0 Step 2: The level of significance is  = Step 3: Using technology, we obtained an estimate of  1 in Parallel Example 2, b 1 = To determine the standard deviation of b 1, we compute. The calculations are on the next slide.

14-28

14-29 Solution Step 3, cont’d: We have The test statistic is

14-30 Solution: P-Value Approach Step 4: Since this is a two-tailed test, the P-value is the sum of the area under the t-distribution with 12- 2=10 degrees of freedom to the left of -t 0 = and to the right of t 0 = Using Table VI we find that with 10 degrees of freedom, the value is between and corresponding to right-tail areas of and 0.001, respectively. Thus, the P-value is between and Step 5: Since the P-value is less than the level of significance, 0.05, we reject the null hypothesis.

14-31 Solution Step 6: There is sufficient evidence at the  = 0.05 level of significance to conclude that a linear relation exists between drill depth and drill time.

14-32 Confidence Intervals for the Slope of the Regression Line A (1-  )  100% confidence interval for the slope of the true regression line,  1, is given by the following formulas: Lower bound: Upper bound: Here, t  /2 is computed using n-2 degrees of freedom.

14-33 Note: The confidence interval formula for  1 can be computed only if the data are randomly obtained, the residuals are normally distributed, and there is constant error variance.

14-34 Parallel Example 7: Constructing a Confidence Interval for the Slope of the True Regression Line Construct a 95% confidence interval for the slope of the least-squares regression line for the drilling example.

14-35 Solution The requirements for the usage of the confidence interval formula were verified in previous examples. We also determined b 1 = in previous examples.

14-36 Solution Since t =2.228 for 10 degrees of freedom, we have Lower bound =  0.003= Upper bound =  0.003= We are 95% confident that the mean increase in the time it takes to drill 5 feet for each additional foot of depth at which the drilling begins is between and minutes.

The Coefficient of Determination The Coefficient of Determination is the proportion of the variability in the response variable that can be attributed to the least squares regression model.

How to calculate R 2 Using the sum of squares technique: But for the SLR models we can simplify the calculation slightly and, where r is the correlation between the response and predictor variables.

14-39 Parallel Example 8: Calculating the Coefficient of Determination Using technology for our drilling example we can calculate the correlation between the response and predictor to be Using the simplified calculation for the coefficient of determination that means:

Interpretation: Our model using the depth at which drilling begins as a predictor is able to explain 59.73% of the natural variability in the time it takes to drill 5 feet.