Presentation is loading. Please wait.

Presentation is loading. Please wait.

Inference for Regression

Similar presentations


Presentation on theme: "Inference for Regression"— Presentation transcript:

1 Inference for Regression
STAT E-150 Statistical Methods Inference for Regression

2 We have discussed how to find a simple linear regression model to make predictions. Now we will investigate our model further: - How do we evaluate the effectiveness of the model? - How do we know that the relationship is significant? - How much of the variability in the response variable can be explained by its relationship to the predictor? The simple linear model is of the form y = β1x + β0 + ε where ε ~ N(0, σε). Recall that inference methods can address questions about the population based on our sample data.

3 Consider our earlier example:
Medical researchers have noted that adolescent females are more likely to deliver low-birthweight babies than are adult females. Because LBW babies tend to have higher mortality rates, studies have been conducted to examine the relationship between birthweight and the mother’s age. We found the fitted regression model

4 How can we assess this model?
If the slope β1 is equal to zero, then there is no change in the response variable associated with a change in the predictor. We will therefore test the value of the slope to investigate whether there is a linear relationship between the two variables

5 T-test for the Slope of a Simple Linear Model H0: β1=0 Ha: β1≠0 The test statistic is with n-2 degrees of freedom.

6 Assumptions for the model and the errors:
1. Linearity Assumption Straight Enough Condition: does the scatterplot appear linear? Check the residuals to see if they appear to be randomly scattered Quantitative Data Condition: Is the data quantitative?

7 Assumptions for the model and the errors:
2. Independence Assumption: the errors must be mutually independent Randomization Condition: the individuals are a random sample Check the residuals for patterns, trends, clumping

8 Assumptions for the model and the errors:
3. Equal Variance Assumption: the variability of y should be about the same for all values of x Does The Plot Thicken? Condition: Is the spread about the line nearly constant in the scatterplot? Check the residuals for any patterns

9 Assumptions for the model and the errors:
4. Normal Population Assumption: the errors follow a Normal model at each value of x Nearly Normal Condition: Look at a histogram or NPP of the residuals

10 We can check the conditions to see if all assumptions are true.
If this is the case, the idealized regression line will have a distribution of y-values for each x-value, and these distributions will be approximately normal with equal variation and with means along the regression line:

11 Here are the steps: Create a scatterplot to see if the data is “straight enough”. Fit a regression model and find the residuals (ε) and predicted values ( ). Draw a scatterplot of the residuals vs. x or ; this should have no pattern or bend, or thickening, or thinning, or outliers. If the scatterplot is “straight enough”, create a histogram and NPP of the residuals to check the “nearly normal” condition. Continue with the inference if all conditions are reasonably satisfied.

12 Here are the results for our data:
- The scatterplot of the data indicates a positive linear relationship between the variables

13 - The scatterplot of the residuals shows no particular pattern

14 - The Normal Probability Plot for the residuals indicates a Normal
- The Normal Probability Plot for the residuals indicates a Normal distribution 

15 Here is our data: H0: Ha: Observation 1 2 3 4 5 6 7 8 9 10
Ha: Observation 1 2 3 4 5 6 7 8 9 10 Maternal Age (in years) 15 17 18 16 19 Birthweight (in grams) 2289 3393 3271 2648 2897 3327 2970 2535 3138 3573

16 Here is our data: H0: β1 = 0 Ha: β1 ≠ 0 Observation 1 2 3 4 5 6 7 8 9
Ha: β1 ≠ 0 Observation 1 2 3 4 5 6 7 8 9 10 Maternal Age (in years) 15 17 18 16 19 Birthweight (in grams) 2289 3393 3271 2648 2897 3327 2970 2535 3138 3573

17 The SPSS output included this table:
What does this value represent?

18 The SPSS output included this table:
What does this value represent?

19 The SPSS output included this table:
What does this value represent? This is the observed (sample) value of the y-intercept

20 The SPSS output included this table:
What does this value represent?

21 The SPSS output included this table:
245.15 What does this value represent?

22 The SPSS output included this table:
245.15 What does this value represent? This is the observed (sample) value of the slope

23 The SPSS output included this table:
What does this value represent?

24 The SPSS output included this table:
45.908 What does this value represent?

25 The SPSS output included this table:
45.908 What does this value represent? This is the standard error for the slope; this is how much we expect the sample slope to vary from one sample to another. 

26 What is the value of the test statistic?

27 5.34 What is the value of the test statistic?

28 5.34 What is the value of the test statistic? 5.34

29 What is the p-value?

30 What is the p-value? p = .001

31 What is the p-value? p = .001 What is your decision?

32 What is the p-value? p = .001 What is your decision? Since p is small, reject the null hypothesis

33 What is the p-value? p = .001 What is your decision? Since p is small, reject the null hypothesis What can you conclude?

34 What is the p-value? p = .001 What is your decision? Since p is small, reject the null hypothesis What can you conclude? The data indicates that there is a linear relationship between the mother’s age and the baby’s birthweight.

35 Confidence Interval for the Slope
The slope of the population regression line, , is the rate of change of the mean response as the explanatory variable increases. The slope of the least squares line, , is an estimate of . A confidence interval for the slope will show how accurate the estimate is.

36 What is the confidence interval for the slope of the regression line?

37 What is the confidence interval for the slope of the regression line?
( , )

38 Confidence Interval for the Slope of a Simple Linear Model
The confidence interval has the form where t* is the critical value for the tn-2 density curve to obtain the desired confidence level. How can we construct this?

39 Confidence Interval for the Slope of a Simple Linear Model
The confidence interval has the form We know that = and that = , but how do we find t*?

40 Confidence Interval for the Slope of a Simple Linear Model
The confidence interval has the form We know that df = n - 2 = 8. On the line for df=8, the value of t for a 95% confidence interval is

41 Calculate = ± 2.306(45.908) = ± = (139.29, )

42 Calculate = ± 2.306(45.908) = ± = (139.29, ) What is the confidence interval found by SPSS?

43 Calculate = ± 2.306(45.908) = ± = (139.29, ) What is the confidence interval found by SPSS?

44 Calculate = ± 2.306(45.908) = ± = (139.29, ) What does this interval tell us?

45 Calculate = ± 2.306(45.908) = ± = (139.29, ) What does this interval tell us? Based on the sample data, we are 95% confident that the true average increase in the weight of the baby associated with a one-year increase in age of the mother is between and g.

46 Partitioning Variability - ANOVA
ANOVA measures the effectiveness of the model by measuring how much of the variability in the response variable y is explained by the predictions based on the fitted model. We can partition this variability into two parts: the variability explained by the model, and the unexplained variability due to error, as measured by the residuals.

47 In our SPSS output, SS(Model) = SS(Error) = SS(Total) =

48 In our SPSS output, SS(Model) = SS(Error) = SS(Total) =

49 What is the value of the test statistic?
What is the p-value? Decision: Conclusion:

50 What is the value of the test statistic? 28.515
What is the p-value? Decision: Conclusion:

51 What is the value of the test statistic? 28.515
What is the p-value? .001 Decision: Conclusion:

52 What is the value of the test statistic? 28.515
What is the p-value? .001 Decision: Since p is small, reject the null hypothesis Conclusion: The data indicates that there is a linear relationship between the mother’s age and the baby’s birthweight.

53 The Coefficient of Determination
This represents the amount of variation in the response variable can be explained by the model: In our example, This tells us that 75.44% of the variation in the birthweights can be explained by the model.

54 Inference for Correlation
There is a relationship between the correlation between the variables and the slope of the least squares line: So testing the value of the slope is the same as testing that there is no correlation between the variables; that is, that the correlation coefficient = 0.

55 T-Test for Correlation
H0: ρ = 0 Ha: ρ ≠ 0 The test statistic is If the conditions of the simple linear model hold, we find the p-value using the t-distribution with n-2 degrees of freedom.

56 The three tests we have discussed are equivalent; it can be shown that the test statistic F is the square of the test statistic t: In our t-test, we found the test statistic t = 5.34 In the ANOVA, we found the test statistic F = t2 = (5.34)2 = = F Also, for both tests we found p = .001

57 Standard Errors for Predicted Values
Let x be a specific value of x. The predicted value of y is We can create two different intervals: a prediction interval for an individual value of x a confidence interval for the mean predicted value at x υ = “nu”

58 The basic format for an interval is
When we want to find a mean predicted value, When we want to find an individual predicted value, Since individual values vary more than means, and the prediction interval will be wider than the confidence interval.


Download ppt "Inference for Regression"

Similar presentations


Ads by Google