MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.

Slides:



Advertisements
Similar presentations
All Possible Regressions and Statistics for Comparing Models
Advertisements

Lesson 10: Linear Regression and Correlation
Kin 304 Regression Linear Regression Least Sum of Squares
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Best subsets regression
3.2 OLS Fitted Values and Residuals -after obtaining OLS estimates, we can then obtain fitted or predicted values for y: -given our actual and predicted.
Chicago Insurance Redlining Example Were insurance companies in Chicago denying insurance in neighborhoods based on race?
Chapter 10 Curve Fitting and Regression Analysis
Objectives (BPS chapter 24)
Chapter 17 Making Sense of Advanced Statistical Procedures in Research Articles.
1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Summarizing Bivariate Data Introduction to Linear Regression.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Applied Linear Regression CSTAT Workshop March 16, 2007 Vince Melfi.
MARE 250 Dr. Jason Turner Multiway, Multivariate, Covariate, ANOVA.
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.
Part I – MULTIVARIATE ANALYSIS C3 Multiple Linear Regression II © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Multiple Regression Involves the use of more than one independent variable. Multivariate analysis involves more than one dependent variable - OMS 633 Adding.
MARE 250 Dr. Jason Turner Multiway, Multivariate, Covariate, ANOVA.
Correlation MARE 250 Dr. Jason Turner.
Lecture 6: Multiple Regression
Multiple Regression MARE 250 Dr. Jason Turner.
Linear Regression MARE 250 Dr. Jason Turner.
REGRESSION AND CORRELATION
MARE 250 Dr. Jason Turner Correlation & Linear Regression.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Chapter 15: Model Building
Correlation and Regression Analysis
Model selection Stepwise regression. Statement of problem A common problem is that there is a large set of candidate predictor variables. (Note: The examples.
1 1 Slide Simple Linear Regression Chapter 14 BA 303 – Spring 2011.
M23- Residuals & Minitab 1  Department of ISM, University of Alabama, ResidualsResiduals A continuation of regression analysis.
Variable selection and model building Part II. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Selecting Variables and Avoiding Pitfalls Chapters 6 and 7.
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
Chapter 14 Multiple Regression Models. 2  A general additive multiple regression model, which relates a dependent variable y to k predictor variables.
Shonda Kuiper Grinnell College. Statistical techniques taught in introductory statistics courses typically have one response variable and one explanatory.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Summarizing Bivariate Data
Chapter 10 Lecture 2 Section: We analyzed paired data with the goal of determining whether there is a linear correlation between two variables.
Chapter 7 Relationships Among Variables What Correlational Research Investigates Understanding the Nature of Correlation Positive Correlation Negative.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Regression Chapter 16. Regression >Builds on Correlation >The difference is a question of prediction versus relation Regression predicts, correlation.
MARE 250 Dr. Jason Turner Multiple Regression. y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept.
Created by Erin Hodgess, Houston, Texas
Aron, Aron, & Coups, Statistics for the Behavioral and Social Sciences: A Brief Course (3e), © 2005 Prentice Hall Chapter 12 Making Sense of Advanced Statistical.
Statistics: Unlocking the Power of Data Lock 5 STAT 101 Dr. Kari Lock Morgan Multiple Regression SECTION 10.3 Variable selection Confounding variables.
Multiple Regression Selecting the Best Equation. Techniques for Selecting the "Best" Regression Equation The best Regression equation is not necessarily.
MARE 250 Dr. Jason Turner Linear Regression. Linear regression investigates and models the linear relationship between a response (Y) and predictor(s)
 Relationship between education level, income, and length of time out of school  Our new regression equation: is the predicted value of the dependent.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Variable Selection 1 Chapter 8 Variable Selection Terry Dielman Applied Regression Analysis:
1 Building the Regression Model –I Selection and Validation KNN Ch. 9 (pp )
Regression. Outline of Today’s Discussion 1.Coefficient of Determination 2.Regression Analysis: Introduction 3.Regression Analysis: SPSS 4.Regression.
Outline of Today’s Discussion 1.Regression Analysis: Introduction 2.An Alternate Formula For Regression 3.Correlation, Regression, and Statistical Significance.
Using SPSS Note: The use of another statistical package such as Minitab is similar to using SPSS.
Variable selection and model building Part I. Statement of situation A common situation is that there is a large set of candidate predictor variables.
Week of March 23 Partial correlations Semipartial correlations
Multiple Independent Variables POLS 300 Butz. Multivariate Analysis Problem with bivariate analysis in nonexperimental designs: –Spuriousness and Causality.
Descriptive measures of the degree of linear association R-squared and correlation.
Canadian Bioinformatics Workshops
Model selection and model building. Model selection Selection of predictor variables.
Stats Methods at IC Lecture 3: Regression.
Theme 6. Linear regression
How regression works The Right Questions about Statistics:
Statistics 200 Lecture #5 Tuesday, September 6, 2016
Kin 304 Regression Linear Regression Least Sum of Squares
Multiple Regression.
Making Sense of Advanced Statistical Procedures in Research Articles
CHAPTER 26: Inference for Regression
Chapter 14 Multiple Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

MARE 250 Dr. Jason Turner Multiple Regression

y Linear Regression y = b 0 + b 1 x y = dependent variable b 0 + b 1 = are constants b 0 = y intercept b 1 = slope x = independent variable Urchin density = b 0 + b 1 (salinity)

Multiple regression allows us to learn more about the relationship between several independent or predictor variables and a dependent or criterion variable For example, we might be looking for a reliable way to estimate the age of AHI at the dock instead of waiting for laboratory analyses Multiple Regression y = b 0 + b 1 x y = b 0 + b 1 x 1 + b 2 x 2 …b n x n

In the social and natural sciences multiple regression procedures are very widely used in research Multiple regression allows the researcher to ask “what is the best predictor of...?” For example, researchers might want to learn what abiotic variables (temp, sal, DO, turb) are the best predictors of plankton abundance/diversity in Hilo Bay Or Which morphometric measurements are the best predictors of fish age Multiple Regression

The general computational problem that needs to be solved in multiple regression analysis is to fit a straight line to a number of points Multiple Regression In the simplest case - one dependent and one independent variable This can be visualized in a scatterplot

A line in a two dimensional or two-variable space is defined by the equation Y=a+b*X The Regression Equation In the multivariate case, when there is more than one independent variable, the regression line cannot be visualized in the two dimensional space, but can be computed rather easily

Two Methods: Best Subset & Stepwise Analysis Best Subsets: Best subsets regression provides information on the fit of several different models, thereby allowing you to select a model based on four distinct statistics Stepwise: Stepwise regression produces a single model based on a single statistic. How To – Multiple Regression

For data sets with a small number of predictors, best subset regression is preferable to stepwise regression because it provides information on more models. For data sets with a large number of predictors (> 32 in Minitab), stepwise regression is preferable. Stepwise…Subsets

S B O P Vars R-Sq R-Sq(adj) C-p S L M P F X X X X X X X X X X X X X X X X Response is Age Best Subsets 1. Simplest model with the highest R 2 wins! 2. Use Mallows’ Cp to break the tie Who decides – YOU!

Stepwise model-building techniques for regression The basic procedures involve: (1) identifying an initial model (2) iteratively "stepping," that is, repeatedly altering the model at the previous step by adding or removing a predictor variable in accordance with the "stepping criteria," (3) terminating the search when stepping is no longer possible given the stepping criteria Stepwise Regression:

For Example… We are interested in predicting values for Y based upon several X’s…Age of AHI based upon SL, BM, OP, PF We run multiple regression and get the equation: Age = SL BM OP PF We then run a STEPWISE regression to determine the best subset of these variables

How does it work… Stepwise Regression: Age versus SL, BM, OP, PF Alpha-to-Enter: 0.15 Alpha-to-Remove: 0.15 Response is Age on 4 predictors, with N = 84 Step Constant BM T-Value P-Value OP T-Value P-Value SL T-Value 1.96 P-Value S R-Sq R-Sq(adj) Mallows C-p Step 1 – BM variable is added Step 2 – OP variable is added Step 3 – SL variable is added

Who Cares? Best Subsets & Stepwise analysis allows you (i.e. – computer) to determine which predictor variables (or combination of) best explain (can be used to predict) Y Much more important as number of predictor variables increase Helps to make better sense of complicated multivariate data

At this point we are still limited to 2-dimensional graphs; although our statistics have become 3- dimensional… However…

There are 3-dimensional graphical techniques to encompass multivariate datasets Don’t Despair Grasshopper… Cool! When do we learn…

“First learn stand, then learn fly. Nature rule, Daniel-san, not mine.” Miyagi Says… All in good time Daniel-san…