Regression. Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression.

Slides:



Advertisements
Similar presentations
 Population multiple regression model  Data for multiple regression  Multiple linear regression model  Confidence intervals and significance tests.
Advertisements

13- 1 Chapter Thirteen McGraw-Hill/Irwin © 2005 The McGraw-Hill Companies, Inc., All Rights Reserved.
Correlation and Linear Regression.
1 SSS II Lecture 1: Correlation and Regression Graduate School 2008/2009 Social Science Statistics II Gwilym Pryce
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
Understanding the General Linear Model
Statistics for the Social Sciences
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Chapter 12 Simple Regression
Chapter 13 Introduction to Linear Regression and Correlation Analysis
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 13 Introduction to Linear Regression and Correlation Analysis.
SIMPLE LINEAR REGRESSION
Linear Regression and Correlation Analysis
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
Correlation and Regression. Correlation What type of relationship exists between the two variables and is the correlation significant? x y Cigarettes.
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
SIMPLE LINEAR REGRESSION
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Correlation and Regression Analysis
Relationships Among Variables
Lecture 5 Correlation and Regression
Example of Simple and Multiple Regression
Correlation and Regression
Lecture 15 Basics of Regression Analysis
Chapter 12 Correlation and Regression Part III: Additional Hypothesis Tests Renee R. Ha, Ph.D. James C. Ha, Ph.D Integrative Statistics for the Social.
SIMPLE LINEAR REGRESSION
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
Introduction to Linear Regression and Correlation Analysis
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Chapter 13: Inference in Regression
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
Simple Linear Regression Models
Anthony Greene1 Correlation The Association Between Variables.
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
1 Chapter 10 Correlation and Regression 10.2 Correlation 10.3 Regression.
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Chapter 10 Correlation and Regression
Multiple Linear Regression. Purpose To analyze the relationship between a single dependent variable and several independent variables.
Chapter 11 Linear Regression Straight Lines, Least-Squares and More Chapter 11A Can you pick out the straight lines and find the least-square?
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Warsaw Summer School 2015, OSU Study Abroad Program Regression.
Correlation & Regression Chapter 15. Correlation It is a statistical technique that is used to measure and describe a relationship between two variables.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Correlation & Regression Analysis
ANOVA, Regression and Multiple Regression March
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Outline of Today’s Discussion
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics Seventh Edition By Brase and Brase Prepared by: Lynn Smith.
Outline of Today’s Discussion 1.Introduction to Correlation 2.An Alternative Formula for the Correlation Coefficient 3.Coefficient of Determination.
Outline of Today’s Discussion 1.Regression Analysis: Introduction 2.An Alternate Formula For Regression 3.Correlation, Regression, and Statistical Significance.
Chi-Square Analyses.
Outline of Today’s Discussion 1.Seeing the big picture in MR: Prediction 2.Starting SPSS on the Different Models: Stepwise versus Hierarchical 3.Interpreting.
Significance Tests for Regression Analysis. A. Testing the Significance of Regression Models The first important significance test is for the regression.
© The McGraw-Hill Companies, Inc., Chapter 10 Correlation and Regression.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Independent Samples ANOVA. Outline of Today’s Discussion 1.Independent Samples ANOVA: A Conceptual Introduction 2.The Equal Variance Assumption 3.Cumulative.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 9 l Simple Linear Regression 9.1 Simple Linear Regression 9.2 Scatter Diagram 9.3 Graphical.
Theme 6. Linear regression
Regression and Correlation
Regression Analysis.
Correlation and Simple Linear Regression
Correlation and Regression
Simple Linear Regression
Regression Analysis.
Prepared by Lee Revere and John Large
Hypothesis testing and Estimation
Presentation transcript:

Regression

Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression Analysis: Excel 5.Independent Predictors

Part 1 Coefficient of Determination

In correlational research Researchers often use the “r-squared” statistic, also called the “coefficient of determination”, to describe the proportion of Y variability explained by X.

Coefficient of Determination What range of values is possible for the coefficient of determination (the r-squared statistic)?

Coefficient of Determination Example: What is the evidence that IQ is heritable?

Coefficient of Determination R-value for the IQ of identical twins reared apart = 0.6. What is the value of r-squared in this case?

Coefficient of Determination So what proportion of the IQ is unexplained (unaccounted for) by genetics?

Coefficient of Determination Different sciences are characterized by the r-squared values that are deemed impressive. (Chemists might r-squared to be > 0.99).

Coefficient of Determination As we have already seen r-squared is the same as “eta-squared”.

Part 2 Regression Analysis Introduction

Correlation is the process of finding a relationship between variables. Regression is the process of finding the best-fitting trend (line) that describes the relationship between variables. So, correlation and regression are very similar!

Regression Analysis Introduction The ‘r’ statistic can be tested for statistical significance! Potential Pop Quiz Question: What two factors determine the critical value (i.e., the number to beat) when we engage in hypothesis testing?

Regression Analysis Introduction DF for Correlation & Regression Here n stands for the number of pairs of scores. Why would this be n-2, rather than the usual n-1?

Regression Analysis Introduction In general, the formula for the degrees of freedom is the number of observations minus the number of parameters estimated. For correlation, we have one estimate for the mean of X, and another estimate for the mean of Y. For regression, we have one estimate for the slope, and another estimate for y intercept.

Regression Analysis Introduction Slope can also be though of as “rise over run”.

Regression Analysis Introduction The “rise” on the ordinate = Y 2 - Y 1. The “run” on the abscissa = X 2 - X 1.

Regression Analysis Introduction “Rise over run” in pictures.

Regression Analysis Introduction Here, the regression is “linear”…

Regression Analysis Introduction Here, the regression is non-linear! What would the equation look like for this trend?

Regression Analysis Introduction Let’s now return to linear regression, and learn how to manually compute the slope and y-intercept. To compute the slope, we need two quantities that we have already learned. These are SP xy (sums of products) and SS x (sums of squares for X)…

Regression Analysis Introduction

Once we have the slope, it’s easy to get the y-intercept!

Part 3 Regression Analysis: SPSS

Regression Analysis: SPSS Later we’ll go to SPSS and get some practice with regression. The steps in SPSS will be Analyze ---> Regression --> Linear. We will place the criterion (i.e., the Y-axis variable) in the “Dependent” box, and the predictor (i.e., the X-axis variable) in the “Independent(s)” box. Click the “Statistics” box, and check “estimates”, “model fit”, and “descriptives”.

Regression Analysis: SPSS The “Coefficients” Section in the SPSS output contains all the info needed for the regression equation, the r statistic, and the evaluation of Ho (retain or reject). The “Coefficients” Section In SPSS Output

Regression Analysis: SPSS The constant is the “b” in, Y = mX + b. Here, b = The “Coefficients” Section In SPSS Output

Regression Analysis: SPSS The slope is the “m” in, Y = mX + b. Here, m = The “Coefficients” Section In SPSS Output

Regression Analysis: SPSS So, our regression equation is, Y = mX + b. or Y = X The “Coefficients” Section In SPSS Output

Regression Analysis: SPSS The r statistic is the standardized coefficient, Beta. r =.705 The “Coefficients” Section In SPSS Output

Regression Analysis: SPSS Lastly, we look at the ‘sig’ value for the predictor, (which is “EDU” in this case) to determine whether predictor (x-axis variable) is significantly correlated with the criterion (y-axis variable). Evaluate Ho: …do we retain or reject? The “Coefficients” Section In SPSS Output

Part 4 Regression Analysis: Excel

Regression Analysis: Excel Correlation and regression are very similar. If we have a significant correlation, the best-fitting regression line is said to have a slope significantly different from zero. Sometimes it is stated that “the slope departs significantly from zero”.

Regression Analysis: Excel Note: A slope can be very modestly different from zero, and still be “statistically significant” if all data points fall very close to the line. In correlation and regression, statistical significance is determined by the strength of the correlation between two variables (the r-value), and NOT by the slope of the regression line. The significance of the r-value, as always, depends on the alpha level, and the df (which is n-2). Take a peak at the r-value table.

Regression Analysis: Excel

Remember: The regression line (equation) can help us predict one score, given another score, but only if there is a significant r-value. The terminology w/b… “the regression line explains (or accounts for)” 42% of the variability in the scores (if r- squared =.42). To “explain” or “account for” does NOT mean “to cause”. Correlation does not imply causation!

Regression Analysis Continued A synonym for regression is prediction! Recall that prediction is one of the four goals of the scientific method. What were the others? A significant correlation implies a significant capacity for prediction, i.e., a prediction that is reliably better than chance!

Regression Analysis Continued The equation for a straight line, again, is: y = mx + B or Criterion = ( slope * Predictor) + Intercept How many “parameters” in a linear equation? How about a quadratic equation?

Part 5 Independent Predictors

So far, we’ve attempted to use regression for prediction. Specifically, we’ve tried to predict one variable Y (called the criterion), using one other variable (called the predictor). Multiple Regression - the process by which one variable Y (called the criterion) is predicted on the basis of more than one variable (say, X1, X2, X3…).

Independent Predictors Here’s the simple case of one predictor variable. The overlap (in gray) indicates the predictive strength.

Independent Predictors If the overlap in the Venn diagram were to grow, the r-value would grow, too!

Independent Prediction Here’s the same thing again… but we’ll call the the predictor variable X 1. Criterion (Y) Variable X 1

Independent Prediction By adding another predictor variable X 2, we could sharpen our predictions. Why? Criterion (Y) Variable X 1 Variable X 2

Independent Prediction Unfortunately, X 1 and X 2 provide some redundant information about Y, so the predictive increase is small. Criterion (Y) Variable X 1 Variable X 2

Independent Prediction Unfortunately, X 1 and X 2 provide some redundant information about Y, so the predictive increase is small. Criterion (Y) Variable X 1 Variable X 2

Independent Prediction By contrast, variable X 3 has no overlap with either X 1 or X 2, so it would add the most new information. Criterion (Y) Variable X 1 Variable X 2 Variable X 3

Independent Prediction In short, since all three predictors provide some unique information, predictions w/b best when using all three. Criterion (Y) Variable X 1 Variable X 2 Variable X 3

Independent Prediction If you wanted to be more parsimonious and use only two of the three, which two would you pick, and why? Criterion (Y) Variable X 1 Variable X 2 Variable X 3

Independent Predictors That was a conceptual introduction to Multiple Regression (predicting Y scores from more than one variable). We will not learn about the computations for multiple regression in this course (but you will if you take the PSYCH 370 course). For our purposes, simply know that predictions improve to the extent that the various predictors are independent of each other.