Intro to Statistics for the Behavioral Sciences PSYC 1900

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

Tests of Significance for Regression & Correlation b* will equal the population parameter of the slope rather thanbecause beta has another meaning with.
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Inference for Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Prediction with multiple variables Statistics for the Social Sciences Psychology 340 Spring 2010.
Quantitative Data Analysis: Hypothesis Testing
Correlation and regression
Cal State Northridge  320 Andrew Ainsworth PhD Regression.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
PSY 307 – Statistics for the Behavioral Sciences
PPA 501 – Analytical Methods in Administration Lecture 8 – Linear Regression and Correlation.
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
PPA 415 – Research Methods in Public Administration
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Interactions in Regression.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 10: Hypothesis Tests for Two Means: Related & Independent Samples.
The Simple Regression Model
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 9: Hypothesis Tests for Means: One Sample.
Intro to Statistics for the Behavioral Sciences PSYC 1900
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
BCOR 1020 Business Statistics
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Multiple Regression Research Methods and Statistics.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Introduction to Regression Analysis, Chapter 13,
Simple Linear Regression Analysis
Relationships Among Variables
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Lecture 5 Correlation and Regression
Example of Simple and Multiple Regression
Active Learning Lecture Slides
Statistics for the Social Sciences Psychology 340 Fall 2013 Tuesday, November 19 Chi-Squared Test of Independence.
Chapter 13: Inference in Regression
Correlation and Linear Regression
Understanding Multivariate Research Berry & Sanders.
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Chapter 15 Correlation and Regression
F OUNDATIONS OF S TATISTICAL I NFERENCE. D EFINITIONS Statistical inference is the process of reaching conclusions about characteristics of an entire.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
Statistics for the Social Sciences Psychology 340 Fall 2013 Correlation and Regression.
Correlation and Linear Regression. Evaluating Relations Between Interval Level Variables Up to now you have learned to evaluate differences between the.
Examining Relationships in Quantitative Research
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Part IV Significantly Different: Using Inferential Statistics
Regression Chapter 16. Regression >Builds on Correlation >The difference is a question of prediction versus relation Regression predicts, correlation.
11 Chapter 12 Quantitative Data Analysis: Hypothesis Testing © 2009 John Wiley & Sons Ltd.
Chapter 16 Data Analysis: Testing for Associations.
MARKETING RESEARCH CHAPTER 18 :Correlation and Regression.
Regression Analysis © 2007 Prentice Hall17-1. © 2007 Prentice Hall17-2 Chapter Outline 1) Correlations 2) Bivariate Regression 3) Statistics Associated.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 12 Testing for Relationships Tests of linear relationships –Correlation 2 continuous.
Correlation & Regression Analysis
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Research Methodology Lecture No :26 (Hypothesis Testing – Relationship)
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Stats Methods at IC Lecture 3: Regression.
Topics: Multiple Regression Analysis (MRA)
Inference about the slope parameter and correlation
REGRESSION G&W p
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
LESSON 24: INFERENCES USING REGRESSION
6-1 Introduction To Empirical Models
Simple Linear Regression
Product moment correlation
Regression Part II.
Presentation transcript:

Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Multiple Regression

Bivariate Regression Review Predicts values of Y as a linear function of X The intercept: a The predicted value of Y when X=0 The slope: b The change in Y associated with a one unit change in X Regression line is one that minimizes errors

Partitioning Variance Total variance of Y is partitioned into portion explained by X and error R2 is portion that is explained Standard error of estimate is average deviation around prediction line

r r2 1 - r2 .10 .01 .99 .20 .04 .96 .30 .09 .91 .40 .16 .84 .50 .25 .75 .60 .36 .64 .70 .49 .51 .80 .90 .81 .19 .95

Hypothesis Testing in Regression The null hypothesis is simply that the slope equals zero. This is equivalent to testing r=0 in correlation. If the correlation is significant, so must the slope be. The actual significance of the slope is tested using a t-distribution. The logic is similar to all hypothesis testing. We compare the magnitude of the slope (b) to its standard error (i.e., the variability of slopes drawn from a population where the null is true).

Hypothesis Testing in Regression The formula to calculate the t value is: Note that standard error of b increases as standard error of the estimate increases We then determine how likely it would be that we found a slope as large as we did using a t distribution (similar to the normal distribution).

Multiple Regression Allows analysis of more than one independent variable Explains variance in Y as a function of a linear composite of iv’s Each iv has a regression coefficient which provides an estimate of its independent effects.

An Example Suppose we want to examine two factors related to the percentage of low birth weights in an area. Percentage of mothers with low education levels Percentage of mothers not seeking prenatal care

Looking at Correlations Here it appears that both low education and lack of care predict increased lower birth weights. Is it fair to say that both variables might be causes?

Example Continued Given that low education and lack of prenatal care are associated, we need to control for one while examining the relation of the other to low birth weights. Multiple regression accomplishes this task.

Another Example Let’s examine applicant attractiveness as a function of GREV, Letters of Rec, & Personal Statements Y, letters, & statements are rated on 1-7 scales If all are significant, which variables are “stronger” predictors? Use standardized regression coefficients

Standardized Regressions The use of standardized coefficients allows easier comparisons of the magnitude of effects Coefficients refer to changes in predicted z-scores of Y as a function of z-scores of x What is the relation of b to r here? In multiple regression, b only equals r if all iv’s are uncorrelated.

Squared Multiple Correlation Coefficient In bivariate regression, we saw that r2 provided an estimate of the variance in Y explained by variance in X. In multiple regression, R2 serves this value. Note that capital R refers to the correlation of the linear composite (i.e., the weighted combination of all iv’s) with Y.

Squared Multiple Correlation Coefficient Returning to the low birth weight example: Here we see that education level and prenatal care explain 67% of the variance in the percentage of low birth weights.

Hypothesis Testing There are two types of null hypotheses in multiple regression. Null for the entire model The linear function (i.e., all variables) does not explain any variance in Y. R2 = 0 If you cannot reject this null, the analysis is over. Null for each predictor Any given iv does not explain variance in Y. b = 0

Testing Hypotheses for Individual Predictors Hypothesis testing here is quite similar to that for the single iv in bivariate regression. But note that in multiple regression, the standard error is sensitive to the overlap (i.e., correlation) among coefficients. As the intercorrelation increases, so does the standard error.

Testing Hypotheses for the Model Here, we must determine if R2 differs form zero. This is accomplished using the F-distribution (we’ll see this again in ANOVA). Logic is the same, however. Is R2 in the sample large enough to make the null unlikely in the population? p=.02

Refining a Model When you are building a model, one way to determine if the addition of new variables improves fit is to see if adding them results in a significant change in R2. If so, this means that adding the variable explains a significant amount of previously unexplained variability. Why not just add variables that significantly correlate with Y?

Time for Examples