Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.

Slides:



Advertisements
Similar presentations
Managerial Economics in a Global Economy
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Bivariate Analyses.
Chapter 10 Curve Fitting and Regression Analysis
Describing Relationships Using Correlation and Regression
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
Cal State Northridge  320 Andrew Ainsworth PhD Regression.
Chapter 13 Multiple Regression
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Multiple regression analysis
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Chapter 12 Simple Regression
Chapter 12 Multiple Regression
Intro to Statistics for the Behavioral Sciences PSYC 1900
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Ch. 14: The Multiple Regression Model building
Chapter 9 - Lecture 2 Computing the analysis of variance for simple experiments (single factor, unrelated groups experiments).
Basic Statistical Concepts Part II Psych 231: Research Methods in Psychology.
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Correlation and Regression Analysis
Simple Linear Regression Analysis
Multiple Regression Models
Relationships Among Variables
Correlation and Regression
Correlation and Linear Regression
Example of Simple and Multiple Regression
Lecture 16 Correlation and Coefficient of Correlation
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Introduction to Linear Regression and Correlation Analysis
Elements of Multiple Regression Analysis: Two Independent Variables Yong Sept
Regression with 2 IVs Generalization of Regression from 1 to 2 Independent Variables.
Chapter 15 Correlation and Regression
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Correlation and Regression Used when we are interested in the relationship between two variables. NOT the differences between means or medians of different.
Multiple regression - Inference for multiple regression - A case study IPS chapters 11.1 and 11.2 © 2006 W.H. Freeman and Company.
MBP1010H – Lecture 4: March 26, Multiple regression 2.Survival analysis Reading: Introduction to the Practice of Statistics: Chapters 2, 10 and 11.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
1 Regression & Correlation (1) 1.A relationship between 2 variables X and Y 2.The relationship seen as a straight line 3.Two problems 4.How can we tell.
Environmental Modeling Basic Testing Methods - Statistics III.
Midterm Review Ch 7-8. Requests for Help by Chapter.
Statistics in IB Biology Error bars, standard deviation, t-test and more.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Analysis of Variance 11/6. Comparing Several Groups Do the group means differ? Naive approach – Independent-samples t-tests of all pairs – Each test doesn't.
Chapter 13 Understanding research results: statistical inference.
Copyright © 2008 by Nelson, a division of Thomson Canada Limited Chapter 18 Part 5 Analysis and Interpretation of Data DIFFERENCES BETWEEN GROUPS AND RELATIONSHIPS.
INTRODUCTION TO MULTIPLE REGRESSION MULTIPLE REGRESSION MODEL 11.2 MULTIPLE COEFFICIENT OF DETERMINATION 11.3 MODEL ASSUMPTIONS 11.4 TEST OF SIGNIFICANCE.
Multiple Regression.
Chapter 14 Introduction to Multiple Regression
Chapter 20 Linear and Multiple Regression
REGRESSION G&W p
Review Guess the correlation
Regression 11/6.
Regression 10/29.
Inference for Regression
Regression.
Correlation and Regression
CHAPTER 29: Multiple Regression*
Multiple Regression.
Introduction to Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9

Review Calculate the correlation between X and Y. z X = [ ] z Y = [ ] A B.-.30 C.-.24 D.-.01

Review Which statement is true? A.If the correlation is zero, then the variables are independent B.If the correlation is nonzero, then the variables are independent C.If two variables are independent, then their correlation is zero D.If two variables have a nonlinear relationship, then their correlation is zero

Regression 11/4

Regression Correlation can tell how to predict one variable from another What about multiple predictor variables? – Explaining Y using X 1, X 2, … X m – Income based on age, education – Memory based on pre-test, IQ – Social behavior based on personality dimensions Regression – Finds best combination of predictors to explain outcome variable – Determines unique contribution of each predictor – Can test whether each predictor has reliable influence

Linear Prediction One predictor – Draw a line through data Multiple predictors – Each predictor has linear effect – Effects of different predictors simply add together Intercept: b 0 – Value of Y when all Xs are zero Regression coefficients: b i – Influence of X i – Sign tells direction; magnitude tells strength – Can be any value (not standardized like correlation)

Example Predict income from education and intelligence – Y = $k/yr – X 1 = years past HS – X 2 = IQ Regression equation: 4-year college, IQ = *4 + 1*120 = 90  $90k/yr 2-year college, IQ = *2 + 1*90 = 40  $40k/yr

Finding Regression Coefficients Goal: regression equation with best fit to data – Minimize squared error between Y and Math solution – Matrix algebra – Combines Xs into a matrix Rows for subjects, columns for variables – Transposes, matrix inverse, derivatives and other fun Practical solution – Use a computer – R: linear model function, lm(Y~X1+X2+X3)

Regression vs. Correlation One predictor: regression = correlation – Plus commitment about which variable predicts the other – And not standardized Multiple predictors – Predictors affect each other – Each coefficient depends on what other predictors are included Example: 3 rd -variable problem – Math achievement and verbal achievement both depend on overall IQ – Math helps with English? – Add IQ as a predictor: verbal ~ math + IQ – b math will be near 0 – No effect of math once IQ is accounted for Regression finds best combination of predictors to explain outcome – Best values of b 1, b 2, etc. taken together – Each coefficient shows contribution of predictor beyond effects of others – Allows identification of most important predictors

Explained Variability Total sum of squares – Variability in Y – SS Y =  (Y – M Y ) 2 Residual sum of squares – Deviation between predicted and actual scores – SS residual =  (Y – ) 2

Explained Variability SS regression – Variability explained by regression – Difference between total and residual sums of squares – SS regression = SS Y - SS residual R 2 ("R-squared") – Fraction of variability explained by the regression – SS regression /SS Y – Measures how well the predictors (Xs) can predict or explain the outcome (Y) – Same as r 2 from correlation when only one predictor SS Y SS residual SS regression

Hypothesis Testing with Regression Does the regression explain meaningful variance in the outcome? Null hypothesis – All regression coefficients equal zero in population – Regression only “explaining” random error in sample Approach – Find how much variance the predictors explain: SS regression – Compare to amount expected by chance Ratio gives F statistic, which we can use for hypothesis testing – If F is larger than expected by H 0, regression explains more variance than expected by chance – Reject null hypothesis if F is large enough

Hypothesis Testing with Regression Likelihood function for F is an F distribution – Comes from ratio of two chi-square variables – Two df values, from numerator and denominator SS Y SS residual SS regression P(F 5,10 ) F crit  F p

Degrees of Freedom SS Y SS regression df Y = n - 1 df regression = m df residual = n – m – 1

Testing Individual Predictors Does predictor variable have any effect on outcome? – Does it provide any information beyond other predictors? – Null hypothesis: b i = 0 in population – Other bs might be nonzero Compute standard error for b i – Uses MS residual to estimate t statistic: Get t crit or p-value from t distribution with df residual Can do one-tailed test if predicted direction of effect – Sign of b i

Review Predicting income from education and IQ: A high-school graduate with IQ = 125 considers getting a 4-year bachelor’s degree. What is her expected increase in income? A.$4k B.$40k C.$95k D.$165k

Review A study investigating personality differences in speech rate measures 100 subjects on 5 personality dimensions. These dimensions are used to predict words per minute (WPM) in a timed speech. The total variability in the outcome variable is SS WPM = The residual variability after accounting for the personality predictors is SS residual = What is R 2 for the regression? A.0.06 B.0.25 C.0.75 D.30 E.3000

Review A study investigating personality differences in speech rate measures 100 subjects on 5 personality dimensions. These dimensions are used to predict words per minute (WPM) in a timed speech. The total variability in the outcome variable is SS WPM = The residual variability after accounting for the personality predictors is SS residual = Calculate the F statistic for testing whether personality predicts speech rate. (df regression = 5, df residual = 94) A.0.02 B.0.33 C.4.70 D.6.27