1 Psych 5510/6510 Chapter Eight--Multiple Regression: Models with Multiple Continuous Predictors Part 3: Testing the Addition of Several Parameters at.

Slides:



Advertisements
Similar presentations
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
Advertisements

Prediction with multiple variables Statistics for the Social Sciences Psychology 340 Spring 2010.
Chapter 17 Making Sense of Advanced Statistical Procedures in Research Articles.
Cal State Northridge  320 Andrew Ainsworth PhD Regression.
Statistics for the Social Sciences Psychology 340 Spring 2005 Prediction cont.
Statistics for the Social Sciences
Chapter 12 Simple Regression
Chapter 11 One-way ANOVA: Models with a Single Categorical Predictor
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
Statistical Methods Chichang Jou Tamkang University.
Lecture 11 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
Lecture 22 Multiple Regression (Sections )
1 Psych 5510/6510 Chapter Eight--Multiple Regression: Models with Multiple Continuous Predictors Part 2: Testing the Addition of One Parameter at a Time.
Simple Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas –quantitative vs. binary predictor.
Lecture 6: Multiple Regression
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Biol 500: basic statistics
Multiple Linear Regression
Statistics for the Social Sciences Psychology 340 Spring 2005 Hypothesis testing with Correlation and Regression.
Chapter 15: Model Building
Multiple Regression – Basic Relationships
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Basic Relationships Purpose of multiple regression Different types of multiple regression.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Split Sample Validation General criteria for split sample validation Sample problems.
Checking Regression Model Assumptions NBA 2013/14 Player Heights and Weights.
Chapter 13: Inference in Regression
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved OPIM 303-Lecture #9 Jose M. Cruz Assistant Professor.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
1 1 Slide © 2012 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 1 Slide Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination n Model Assumptions n Testing.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
1 Psych 5510/6510 Chapter 12 Factorial ANOVA: Models with Multiple Categorical Predictors and Product Terms. Spring, 2009.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
1 Psych 5510/6510 Chapter 10. Interactions and Polynomial Regression: Models with Products of Continuous Predictors Spring, 2009.
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Examining Relationships in Quantitative Research
MBP1010H – Lecture 4: March 26, Multiple regression 2.Survival analysis Reading: Introduction to the Practice of Statistics: Chapters 2, 10 and 11.
Correlation & Regression Chapter 5 Correlation: Do you have a relationship? Between two Quantitative Variables (measured on Same Person) (1) If you have.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 4 Section 2 – Slide 1 of 20 Chapter 4 Section 2 Least-Squares Regression.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Regression Chapter 16. Regression >Builds on Correlation >The difference is a question of prediction versus relation Regression predicts, correlation.
SW388R6 Data Analysis and Computers I Slide 1 Multiple Regression Key Points about Multiple Regression Sample Homework Problem Solving the Problem with.
Chapter 13 Multiple Regression
1 Psych 5510/6510 Chapter 14 Repeated Measures ANOVA: Models with Nonindependent ERRORs Part 3: Factorial Designs Spring, 2009.
1 Psych 5500/6500 Measures of Variability Fall, 2008.
1 Psych 5510/6510 Chapter Eight--Multiple Regression: Models with Multiple Continuous Predictors Part 1: Testing the Overall Model Spring, 2009.
1 Psych 5510/6510 Chapter 13 ANCOVA: Models with Continuous and Categorical Predictors Part 2: Controlling for Confounding Variables Spring, 2009.
1 Psych 5510/6510 Chapter 14 Repeated Measures ANOVA: Models with Nonindependent Errors Part 1 (Crossed Designs) Spring, 2009.
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 12 Multiple.
ANOVA, Regression and Multiple Regression March
1 Psych 5510/6510 Chapter 13: ANCOVA: Models with Continuous and Categorical Predictors Part 3: Within a Correlational Design Spring, 2009.
Regression. Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression.
1 Psych 5510/6510 Chapter 14 Repeated Measures ANOVA: Models with Nonindependent ERRORs Part 2 (Crossed Designs) Spring, 2009.
More on regression Petter Mostad More on indicator variables If an independent variable is an indicator variable, cases where it is 1 will.
(Slides not created solely by me – the internet is a wonderful tool) SW388R7 Data Analysis & Compute rs II Slide 1.
Choosing and using your statistic. Steps of hypothesis testing 1. Establish the null hypothesis, H 0. 2.Establish the alternate hypothesis: H 1. 3.Decide.
Multiple Regression Scott Hudson January 24, 2011.
BNAD 276: Statistical Inference in Management Spring 2016 Green sheets.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
The simple linear regression model and parameter estimation
Statistics for the Social Sciences
STA 282 Introduction to Statistics
Business Statistics Multiple Regression This lecture flows well with
Prepared by Lee Revere and John Large
Statistics for the Social Sciences
Multiple Regression – Split Sample Validation
Regression Analysis.
Introduction to Regression
Presentation transcript:

1 Psych 5510/6510 Chapter Eight--Multiple Regression: Models with Multiple Continuous Predictors Part 3: Testing the Addition of Several Parameters at a Time Spring, 2009

2 Testing the Addition of a Set of Predictors So far we have looked at: 1.Testing the overall model. 2.Testing the addition of one more parameter to the model. Now we will look at adding more than one parameter at a time (i.e. adding a ‘set of predictors’).

3 Example Let’s say we are interested in modeling voting behavior (Y). We have several psychological measures (e.g. personality and attitude; X 1 and X 2 ) we could use in the model, and we have several socio-demographic measures (e.g. age, income, education; X 3, X 4 and X 5 ), we could use. We start with the set of two psychological measures, find that they are worthwhile, and then see if it is worthwhile to add the set of three socio-demographic variables.

4 Example (cont.) MODEL C: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i + β 4 X 4i + β 5 X 5i H 0 : β 3 = β 4 = β 5 =0 H A : At least one of those 3 betas doesn’t equal 0.

5 Testing H0 Most computer programs—including SPSS—can’t directly test this scenario, as for them Model C is always the 1-parameter model of Ŷ i =β 0.

6 Solution First have SPSS do a linear regression analysis on your Model C: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i SPSS will analyze this as if it were model A in the following setup: MODEL C: Ŷ i =β 0 MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i You do this just to get the SSE for the three-parameter model (which SPSS will report as ‘SS Residual’).

7 Solution Second have SPSS do a linear regression analysis on your six parameter Model A SPSS will analyze this as if it were model A in the following setup: MODEL C: Ŷ i =β 0 MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i + β 4 X 4i + β 5 X 5i You do this just to get the SSE for the six parameter model, which again SPSS will report as SS Residual.

8 Solution Third, from these two analyses you have the SSE’s you want: SSE(C) is the SS residual from the SPSS analysis of the three parameter model. SSE(A) is the SS residual from the SPSS analysis of the six parameter model. Now you can compute SSR, PRE, and then test for the significance using any one of the three methods we have covered (test the PRE for significance, transform the PRE into an F value, or transform the MS’s into an F value).

9 Example We want to predict college GPA, so we begin with the following: MODEL C: Ŷ i =β 0 MODEL A: Ŷ i =β 0 + β 1 X 1i Where Y is college GPA and X 1 is percentile rank in high school.

10 Example (cont.) Let’s say we find that Model A is worthwhile, HS rank significantly improves the model compared to just using the mean of Y. Now we want to know if it is worthwhile to include the two SAT measures (math and verbal). If we have one we also have the other so we might want to include both. If the two are highly redundant, however, then we don’t want to add both, as our PRE per parameter added would drop compared to adding just one. But, on the other hand, they shouldn’t be too redundant or the SAT wouldn’t bother to report both.

11 Example (cont.) What we want to test: MODEL C: Ŷ i =β 0 + β 1 X 1i MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i To get the SSE(C) you want, use SPSS to regress Y on X1. This is your Model C (above) but when you have SPSS do the regression SPSS treats it as Model A in the following setup: MODEL C: Ŷ i =β 0 MODEL A: Ŷ i =β 0 + β 1 X 1i SSE = (from SPSS ‘SS Residual’), this provides the SSE(C) for your analysis. See next slide.

12 ANOVA from regressing Y (GPA) on High School Rank SS residual is from the model: Ŷ i =β 0 + β 1 X Rank

13 Example (cont.) What we want to test: MODEL C: Ŷ i =β 0 + β 1 X 1i MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i To get the SSE(A) you want, set up an analysis where model A is compared to the simple model: MODEL C: Ŷ i =β 0 MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i SSE = (from SPSS ‘SS Residual’) Which is your SSE(A). See next slide.

14 ANOVA from regressing Y (GPA) on High School Rank, SAT Verbal and SAT Math. SS residual is from the model: Ŷ i =β 0 + β 1 X Rank + β 2 X SAT_V + β 3 X SAT_M

15 Example (cont.) MODEL C: Ŷ i =β 0 + β 1 X 1i MODEL A: Ŷ i =β 0 + β 1 X 1i + β 2 X 2i + β 3 X 3i SSE(C)= SSE(A)=172.99, SSR= =16.12 PRE=16.22/189.11=.085 PA=4, PC=2, N=414, PRE critical .012 Or, using the PRE tool, p<.001 Reject H0, adding the two SAT scores is worthwhile.

16 Cautions About Using Multiple Regression 1)Causal Conclusions. Be very cautious about drawing causal conclusions when you don’t manipulate the independent variable(s). That the X variables can be used to predict the Y variable does not prove that X causes Y.

17 2.Judging the Relative Importance of Predictor Variables It is tempting to judge the relative importance of the various predictor variables by comparing the value of their betas. For example, if in the regression formula you have …0.25(X 1 ) (X 2 )… then it is tempting to conclude that X 2 must play a more important role in the model of Y than does X 1. This is incorrect as the value of the beta is influenced by the scale of the predictor variable (e.g. changing X 1 from feet to inches will increase the value of its beta).

18 A way to remove the effect of the scale of X on the value of its beta is to used the standardized betas (which are the betas you use when you change all of your variables to standard scores). That may still not allow you to determine the relative importance of the predictor variables: a)If the researcher decides to use a different range of cases then the standardized regression correlations could change quite a bit even though the relationship between x and y remains the same, so in that regard they are not ‘standard’. b)The size of the standardized regression correlations are still affected by redundancy.

19 3.Automatic Model Building Example: stepwise regression. The predictor variable with highest resulting F* is added (as long as it meets some minimum threshold), then on subsequent steps the remaining variable with the highest F* is added (if it meets the threshold), or a variable already in the model whose F* drops below the threshold (due to redundancy) is removed. Continue until all the variables in the model have F*s above the threshold, and all the variables not in the model have F*s below the threshold.

20 The authors argue against automatic model building: 1.A search unfocused by theory is likely to be a fishing expedition that finds spurious relationships and creates models that don’t replicate with new data. 2.The interpretation of coefficients and the meaning of the question being asked depends upon the other variables in the model, it’s unwise to let an automatic procedure determine what questions we do and do not ask of our data. 3.Better models and better understanding of data result from focused data analysis guided by substantive theory.