Download presentation

Presentation is loading. Please wait.

Published byRoxanne Woods Modified over 4 years ago

1
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression

2
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. 13.1 Aptness of the Model and Model Checking

3
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Standardized Residuals The standardized residuals are given by

4
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Diagnostic Plots The basic plots for an assessment of model validity and usefulness are 1. e i * (or e i ) on the vertical axis vs. x i on the horizontal axis. 2. e i * (or e i ) on the vertical axis vs. y i on the horizontal axis. (these two plots are called residual plots)

5
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Diagnostic Plots 3. on the vertical axis vs. y i on the horizontal axis. 4. A normal probability plot of the standardized residuals.

6
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Difficulties in the Plots 1. A nonlinear probabilistic relationship between x and y is appropriate. 2. The variance of (and of Y) is not a constant but depends on x. 3. The selected model fits well except for a few outlying data values, which may have greatly influenced the choice of the best- fit function.

7
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Difficulties in the Plots 4. The error term does not have a normal distribution. 5. When the subscript i indicates the time order of the observations, the exhibit dependence over time. 6. One or more relevant independent variables have been omitted from the model.

8
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Non linear relationship (use nonlinear model) Non-constant variance (weighted least squares) Discrepant observation Observation with large influence (omit value or MAD) Abnormality in Data (and Remedies)

9
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Abnormality in Data (and Remedies) Dependence in errors (transform y’s or model including time) Variable omitted (multiple regression model including omitted variable)

10
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. 13.2 Regression With Transformed Variables

11
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Intrinsically Linear – Function A function relating y to x is intrinsically linear if by means of a transformation on x and/or y, the function can be expressed as where are the transformed independent and dependent variables respectively.

12
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Intrinsically Linear Functions Function Transformation Linear Form (exponential) (power) (reciprocal)

13
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Intrinsically Linear – Probabilistic Model A probabilistic model relating Y to x is intrinsically linear if by means of a transformation on Y and/or x, the it can be reduced to a linear probabilistic model

14
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Intrinsically Linear Probabilistic Models Exponential Power Reciprocal

15
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Analyzing Transformed Data 1. Estimating and then transforming back to obtain estimates of the original parameters is not equivalent to using the principle of least squares on the original model. 2. If the chosen model is not intrinsically linear, least squares would have to be applied to the untransformed model.

16
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Analyzing Transformed Data 3. When a transformed model satisfies the assumptions in chap.12, the method of least squares yields the best estimates of the transformed parameters. The estimates of the original parameters may not be the best. 4. After a transformation on y, to use standard formulas to test hypotheses or construct CI’s, should be at least normally distributed.

17
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Analyzing Transformed Data 5. When y is transformed, the r 2 value from the resulting regression refers to variation in the explained by the transformation regression model.

18
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Logit Function An example where we have a function of

19
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Logistic Regression Logistic regression means assuming that p(x) is related to x by the logit function. Algebra yields: Called the odds ratio

20
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. 13.3 Polynomial Regression

21
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Polynomial Regression Model The kth-degree polynomial regression model equation is is normally distributed with

22
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Regression Models Quadratic ModelCubic Model

23
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Estimation of Parameters Using Least Squares k + 1 normal equations: Solve for estimates of

24
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Coefficient of multiple determination: R 2 Estimate of Adjusted R 2

25
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Confidence Interval and Test coefficient of x i in the polynomial regression function is A test of is based on the following t statistic value and n – (k + 1) df.

26
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. CI for Let x* denote a particular value of x. with becomes

27
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. PI

28
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Centering x Values Let = the average of the xi’s for which observations are to be taken and consider In this model and the parameters describe the behavior of the regression near the center of the data.

29
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. 13.4 Multiple Regression Analysis

30
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. General Additive Multiple Regression Model Equation For purposes of testing hypotheses and calculating CIs or PIs, assume is normally distributed. where

31
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Models 1. The first-order model: 2. Second-order no-interaction model: 3. First-order predictors and interaction: 4. Second-order (full quadratic) model:

32
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Models With Predictors for Categorical Variables Using simple numerical coding, qualitative (categorical) variables can be incorporated into a model. With a dichotomous variable associate an indicator (dummy) variable x whose possible values are 0 and 1.

33
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Estimation of Parameters Normal equations: Solve for estimates of

34
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Coefficient of multiple determination: R 2 Estimate of Adjusted R 2

35
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Test statistic: Alt. hypoth.: Rejection region: Null hypoth.: Test

36
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. coefficient of x i in the regression function is Inferences Based on the Model Simultaneous CIs for several for which the simultaneous confidence level is controlled can be obtained by the Bonferroni technique.

37
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. 2. A test of is based on the following t statistic value and n – (k + 1) df. The test is upper-, lower-, or two- tailed according to whether H a contains the inequality >, < or Inferences Based on the Model

38
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. where Inferences Based on the Model

39
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Inferences Based on the Model

40
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. An F Test for a Group of Predictors (so is correct) versus

41
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Test statistic: Rejection Region: Procedure SSE k = unexplained variation (full model) SSE l = unexplained variation (reduced model)

42
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. 13.5 Other Issues in Multiple Regression

43
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Transformations in Multiple Regression Theoretical considerations or diagnostic plots may suggest a nonlinear relation between a dependent variable and two or more independent variables. Frequently a transformation will linearize the model.

44
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Standardizing Variables be the sampled average and standard deviation of the x ij ’s, then the coded full second-order model with two independent variables has regression function increased numerical accuracy more accurate estimation for the parameters

45
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Variable Selection 1. If we examine regressions involving all possible subsets of the predictors for which data is available, what criteria should be used to select a model? 2. If the number of predictors is too large to examine all regressions, can a good model be found by examining a reduced number of subsets?

46
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Criteria for Variable Selection the coefficient of multiple determination for a k-predictor model. Identify k for which is nearly as large as R 2 for all predictors in the model. 2. MSE k = SSE/(n – k – 1), the mean squared error for a k –predictor model. Find the model having minimum MSE k.

47
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Criteria for Variable Selection A desirable model is specified by a subset of predictors for which C k is small.

48
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Stepwise Regression When the number of predictors is too large to allow for explicit or implicit examination of all possible subsets, alternative selection procedures generally identify good models. Two of these methods are the backward elimination (BE) method and the forward selection method (FS).

49
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Identifying Influential Observations In general, the predicted values corresponding to the sample observations can be written If h jj > 2(k + 1)/n, the jth observation is potentially influential (some regard 3(k + 1)/n as the value.)

Similar presentations

© 2019 SlidePlayer.com Inc.

All rights reserved.

To make this website work, we log user data and share it with processors. To use this website, you must agree to our Privacy Policy, including cookie policy.

Ads by Google