Partial and Semipartial Correlation

Slides:



Advertisements
Similar presentations
Dummy Dependent variable Models
Advertisements

Things to do in Lecture 1 Outline basic concepts of causality
Cause (Part II) - Causal Systems I. The Logic of Multiple Relationships II. Multiple Correlation Topics: III. Multiple Regression IV. Path Analysis.
Redundancy and Suppression
1 G Lect 4M Interpreting multiple regression weights: suppression and spuriousness. Partial and semi-partial correlations Multiple regression in.
Simple Linear Regression 1. 2 I want to start this section with a story. Imagine we take everyone in the class and line them up from shortest to tallest.
ANCOVA Workings of ANOVA & ANCOVA ANCOVA, Semi-Partial correlations, statistical control Using model plotting to think about ANCOVA & Statistical control.
Linear Regression.  The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:  The model won’t be perfect, regardless.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Instrumental Variables Estimation and Two Stage Least Square
Regression Basics Predicting a DV with a Single IV.
Lecture 3 Cameron Kaplan
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Multiple Regression Models: Some Details & Surprises Review of raw & standardized models Differences between r, b & β Bivariate & Multivariate patterns.
1 Psych 5510/6510 Chapter Eight--Multiple Regression: Models with Multiple Continuous Predictors Part 2: Testing the Addition of One Parameter at a Time.
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
Multiple Regression 2 Sociology 5811 Lecture 23 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Relationships Among Variables
Structural Equation Models – Path Analysis
Example of Simple and Multiple Regression
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
CORRELATION & REGRESSION
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Multiple Regression 1 Sociology 5811 Lecture 22 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
7.1 Multiple Regression More than one explanatory/independent variable This makes a slight change to the interpretation of the coefficients This changes.
Chapter 7 Regression. Difference between correlation and regression Regression (Tendency of regressing to the mean) In correlation there is no distinction.
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Welcome to Econ 420 Applied Regression Analysis Study Guide Week Two Ending Sunday, September 9 (Note: You must go over these slides and complete every.
Soc 3306a Multiple Regression Testing a Model and Interpreting Coefficients.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
Correlation and Regression PS397 Testing and Measurement January 16, 2007 Thanh-Thanh Tieu.
1 G Lect 6M Comparing two coefficients within a regression equation Analysis of sets of variables: partitioning the sums of squares Polynomial curve.
Chapter 9 Analyzing Data Multiple Variables. Basic Directions Review page 180 for basic directions on which way to proceed with your analysis Provides.
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
Correlation & Regression Chapter 5 Correlation: Do you have a relationship? Between two Quantitative Variables (measured on Same Person) (1) If you have.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Ordinary Least Squares Regression.
Multiple Regression Petter Mostad Review: Simple linear regression We define a model where are independent (normally distributed) with equal.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
Educ 200C Wed. Oct 3, Variation What is it? What does it look like in a data set?
Discussion of time series and panel models
Chapter 8 Linear Regression. Slide 8- 2 Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
1 Psych 5510/6510 Chapter Eight--Multiple Regression: Models with Multiple Continuous Predictors Part 1: Testing the Overall Model Spring, 2009.
Overview of Regression Analysis. Conditional Mean We all know what a mean or average is. E.g. The mean annual earnings for year old working males.
CHAPTER 8 Linear Regression. Residuals Slide  The model won’t be perfect, regardless of the line we draw.  Some points will be above the line.
Correlation & Regression Analysis
Correlation They go together like salt and pepper… like oil and vinegar… like bread and butter… etc.
Multiple Regression David A. Kenny January 12, 2014.
Chapter 8 Relationships Among Variables. Outline What correlational research investigates Understanding the nature of correlation What the coefficient.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Multiple Independent Variables POLS 300 Butz. Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality.
Chapter 11 REGRESSION Multiple Regression  Uses  Explanation  Prediction.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Regression Analysis.
Introduction to Regression Analysis
Regression 11/6.
Regression 10/29.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Regression.
Multiple Regression – Part II
Cause (Part II) - Causal Systems
Introduction to Regression
Product moment correlation
Presentation transcript:

Partial and Semipartial Correlation Working With Residuals

Questions Give a concrete example (names of vbls, context) where it makes sense to compute a partial correlation. Why a partial rather than semipartial? Why is the squared semipartial always less than or equal to the squared partial? Give a concrete example where it makes sense to compute a semipartial correlation. Why semi rather than partial? Why is regression more closely related to semipartials than partials? How could you use ordinary regression to compute 3rd order partials?

Partial Correlation People differ in many ways. When one difference is correlated with an outcome, cannot be sure the correlation is not spurious. Would like to hold third variables constant, but cannot manipulate. Can use statistical control. Statistical control is based on residuals. If we regress X2 on X1 and take residuals of X2, this part of X2 will be uncorrelated with X1, so anything X2 resids correlate with will not be explained by X1.

Example of Partials Use SAT to predict grades (HS & College Fresh) HS=.8557+.0043*SAT; F=.9563+.0038*SAT. (HS) (F) Person SAT-V HSGPA FGPA PFGPA E1 E2 1 500 3.0 2.8 2.86 -0.01 -0.06 2 550 3.2 3.05 -0.02 -0.05 3 450 2.67 0.01 0.13 4 400 2.5 2.2 2.48 -0.08 -0.28 5 600 3.3 3.24 -0.24 0.06 6 650 3.8 3.43 0.15 -0.13 7 700 3.9 3.5 3.61 0.03 -0.12 8 3.7 0.58 0.65 9 3.4 -0.15 -0.03 10 3.1 2.9 R2 for HS = .76; R2 for F = .62 (fictional data).

Example Partials (2) There are 2 sets of predicted values; one for each GPA, however, they correlate 1.0 with each other, so only 1 is presented. High correlations   SAT-V HS GPA F P E1 (HS) E2 (F) 1 HSGPA .87 FGPA .81 .92 1.00 .00 .50 .45 E2 .37 .58 .74 Note that P and SAT are perfectly correlated. P & SAT do not correlate with E1 or E2 (residuals). A partial correlation; the correlation between the residuals of the two GPAs. The correlation between HS GPA and FGPA holding SAT constant.

The Meaning of Partials The partial is the result of holding constant a third variable via residuals. It estimates what we would get if everyone had same value of 3rd variable, e.g., corr b/t 2 GPAs if all in sample have SAT of 500. Some examples of partials? Control for SES, prior experience, what else?

Computing Partials from Correlations Although you compute partials via residuals, sometimes it is handy to compute them with correlations. Also looking at the formulas is (could be?) informative. Notation. The partial correlation is r12.3 where variable 3 is being partialed from the correlation between 1 and 2. In our example, The partial correlation can be a little or a lot bigger or smaller than the original.

The Order of a Partial If you partial 1 vbl out of a correlation, the resulting partial is called a first order partial correlation. If you partial 2 vbls out of a correlation, the resulting partial is called a second order partial correlation. Can have 3rd, 4th, etc., order partials. Unpartialed (raw) correlations are called zero order correlations because nothing is partialed out. Can use regression to find residuals and compute partial correlations from the residuals, e.g. for r12.34, regress 1 and 2 on both 3 and 4, then compute correlation between 2 sets of residuals.

Partials from Multiple Correlation We can compute squared partial correlations from various R2 values. is the R2 from the regression in which 1 is the DV and 2 and 3 are the Ivs. Alternative (possibly friendlier) notation.

Squared Partials from R2 - Venn Diagrams Here we want the partial correlation Between Y and X1 holding X2 constant. 2. 1. 3. 4.

Exercise – Find a Partial 1 2 3 1 ANX 2 Fam History .20 3 DOC Visit .35 .15 What is the correlation between trait anxiety and the number of doctor visits controlling for family medical history?

Find a partial 1 2 3 1 ANX 2 Fam History .20 3 DOC Visit .35 .15

Semipartial Correlation With partial correlation, we find the correlation between X and Y holding Z constant for both X and Y. Sometimes, we want to hold Z constant for just X or just Y. Instead of holding constant for both, hold for only one, therefore it’s a semipartial correlation instead of a partial. With a semipartial, we find the residuals of X on Z or Y on Z but the other is the original, raw variable. Correlate one raw with one residual. In our example, we found the correlation between E1 (HSGPA) and FGPA to be .45. This is the semipartial correlation between HSGPA and FGPA holding SAT constant for HSGPA only.

Semipartials from Correlations Note that r1(2.3) means the semipartial correlation between variables 1 and 2 where 3 is partialled only from 2. In our example: Agrees with earlier results within rounding error.

Squared Semipartials from Multiple Correlations Squared semipartial is an increment in R2.

Partial vs. Semipartial Why is the squared partial larger than the squared semipartial? Look at the respective areas for Y.

Regression and Semipartial Correlation Regression is essentially about semipartials Each X is residualized on the other X variables. For each X we add to the equation, we ask, “What is the unique contribution of this X above and beyond the others?” Increment in R2 when added last. We do NOT residualize Y, just X. Semipartial because X is residualized but Y is not. b is the slope of Y on X, holding the other X variables constant.

Semipartial and Regression 2 Standardized regression coefficient Semipartial correlation The difference is the square root in the denominator. The regression coefficient can exceed 1.0 in absolute value; the correlation cannot.

Uses of Partial and Semipartial The partial correlation is most often used when some third variable z is a plausible explanation of the correlation between X and Y. Job characteristics and job sat by NA Cog ability and grades by SES The semipartial is most often used when we want to show that some variable adds incremental variance in Y above and beyond other X variable Pilot performance and Cog ability, motor skills Patient well being and surgery, social support

Review Give a concrete example (names of vbls, context) where it makes sense to compute a partial correlation. Why a partial rather than semipartial? Give a concrete example where it makes sense to compute a semipartial correlation. Why semi rather than partial?

Suppressor Effects Hard to understand, but Inspection of r not enough to tell value Need to know to avoid looking dumb Show problems with Venn diagrams Think of observed variable as composite of different stuff, e.g., satisfaction with car (price, prestige, etc.)

Suppressor Effects (2) Note that X2 is correlated with X1 but NOT with Y. Will X2 be useful in a regression equation?   Y X1 X2 1 .50 .00 If we solve for beta weights, we find, beta1=.667 and beta2 = -.333. Notice that the beta weight for the first is actually larger than r (.50), and the second has become negative. Can also happen that r is (usually slightly) positive and beta is negative. This is a suppressor effect. Always inspect your correlations along with your regression weights to see if this is happening. What does it mean that beta2 is negative? Sometimes people forget that there are other X variables in the equation. “The results mean that we should feed people more to get them to lose weight.”

Suppressor Effects (3) Can also happen in path analysis, CSM. Explanation – X2 is a measure of prediction error in X1. If we subtract X2, will have a more useful measure of X1. X2 ‘suppresses’ the correlation of Y and X1. Inspection of correlation matrix not sufficient to see value of variables. Looking dumb. Venn diagram.

Review Why is the squared semipartial always less than or equal to the squared partial? Why is regression more closely related to semipartials than partials? How could you use ordinary regression to compute 3rd order partials?

Exercise – Find a Semipartial   Y X1 X2 1 .20 .30 .40 What is the correlation between Y and X1 holding X2 constant only for X1?

Find a Semipartial The correlation of X1 with Y after controlling for X2 (from X1 only) is rather small.   Y X1 X2 1 .20 .30 .40

Computer Exercise Go to labs and download 2IV Example. Find the partial correlation between hassles and well being holding gender and anger constant (2nd order partial). Find the squared semipartial for anger when well being is the DV and gender and hassles are the other IVs, that is, find the increment in R-square when anger is added to the equation after gender and hassles.