PSY 1950 Regression November 10, 2008. Definition Simple linear regression –Models the linear relationship between one predictor variable and one outcome.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Linear regression models
Simple Regression. Major Questions Given an economic model involving a relationship between two economic variables, how do we go about specifying the.
Definition  Regression Model  Regression Equation Y i =  0 +  1 X i ^ Given a collection of paired data, the regression equation algebraically describes.
Overview Correlation Regression -Definition
Objectives (BPS chapter 24)
Statistics for the Social Sciences
Chapter 10 Simple Regression.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Linear Regression and Linear Prediction Predicting the score on one variable.
1 Chapter 17: Introduction to Regression. 2 Introduction to Linear Regression The Pearson correlation measures the degree to which a set of data points.
Simple Linear Regression Analysis
Linear Regression/Correlation
Relationships Among Variables
Correlation & Regression
Linear Regression.
Introduction to Linear Regression and Correlation Analysis
Simple Linear Regression
Biostatistics Unit 9 – Regression and Correlation.
Managerial Economics Demand Estimation. Scatter Diagram Regression Analysis.
Section 5.2: Linear Regression: Fitting a Line to Bivariate Data.
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
Correlation and Regression Basic Concepts. An Example We can hypothesize that the value of a house increases as its size increases. Said differently,
Simple Linear Regression (OLS). Types of Correlation Positive correlationNegative correlationNo correlation.
Chapter 10 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 A perfect correlation implies the ability to predict one score from another perfectly.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
Simple Linear Regression In the previous lectures, we only focus on one random variable. In many applications, we often work with a pair of variables.
Multiple Regression. Simple Regression in detail Y i = β o + β 1 x i + ε i Where Y => Dependent variable X => Independent variable β o => Model parameter.
Ch14: Linear Least Squares 14.1: INTRO: Fitting a pth-order polynomial will require finding (p+1) coefficients from the data. Thus, a straight line (p=1)
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Psychology 202a Advanced Psychological Statistics October 22, 2015.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
3.2 - Residuals and Least Squares Regression Line.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
Correlation and Regression Basic Concepts. An Example We can hypothesize that the value of a house increases as its size increases. Said differently,
1 AAEC 4302 ADVANCED STATISTICAL METHODS IN AGRICULTURAL RESEARCH Part II: Theory and Estimation of Regression Models Chapter 5: Simple Regression Theory.
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Bivariate Regression. Bivariate Regression analyzes the relationship between two variables. Bivariate Regression analyzes the relationship between two.
Multiple Regression.
The simple linear regression model and parameter estimation
Regression and Correlation of Data Summary
Chapter 20 Linear and Multiple Regression
REGRESSION G&W p
Regression Chapter 6 I Introduction to Regression
3.1 Examples of Demand Functions
Kin 304 Regression Linear Regression Least Sum of Squares
BPK 304W Regression Linear Regression Least Sum of Squares
BPK 304W Correlation.
Simple Linear Regression - Introduction
Correlation and Simple Linear Regression
Multiple Regression.
Linear Regression/Correlation
Regression Models - Introduction
Correlation and Simple Linear Regression
Simple Linear Regression
Simple Linear Regression
Least-Squares Regression
Simple Linear Regression and Correlation
Introduction to Regression
Regression Models - Introduction
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

PSY 1950 Regression November 10, 2008

Definition Simple linear regression –Models the linear relationship between one predictor variable and one outcome variable –e.g., predicting income based upon age Multiple linear regression –Models the linear relationship between more than one predictor variables and one outcome variable –e.g., predicting income based upon age and sex Lingo –Independent/dependent, predictor/outcome

History Astronomical predictions: method of least squares –Piazzi (1801) spotted Ceres, made 22 observations over 41 days, got sick, lost Ceres –Gauss: "... for it is now clearly shown that the orbit of a heavenly body may be determined quite nearly from good observations embracing only a few days; and this without any hypothetical assumption.” Genetics: Regression to the mean –Galton, F. (1886). Regression towards mediocrity in hereditary stature. Journal of the Anthropological Institute, 15, 246–263.

Lines Mathematically, a line is defined by its slope and intercept –Slope is change in Y per change in X –Intercept is the points at which the line crosses the Y-axis, i.e., Y when X = 0 Y = bX + a –b is slope –a is intercept

Which Lines is Best?

Residuals Residuals are –Errors in prediction –Difference between expected values (under your model) and observed values (in your dataset)

Minimizing Residuals Can define the best fit line by summing –Absolute residuals (Method of Least Absolute Deviations) –Squared residuals (Method of Least Squares)

Which is Better? Method of Least Squares –Not robust –Stable (line doesn’t “jump” with small changes in X) –Only one solution (unique line for each dataset) Method of Least Absolute Deviations –Robust –Unstable (line does “jump” with small changes in X) –Multiple solutions (sometimes) ls/SAS/lablets/7.3/7.3c/lab73c.htmlhttp:// ls/SAS/lablets/7.3/7.3c/lab73c.html

Multiple Solutions Any line within the “green zone” produces the same summed residuals via the method of least absolute deviations

Method of (Ordinary) Least Squares

Regression Coefficients Slope Intercept

Standardized Coefficients ^ ^

Regression Line Passes Through (M X, M Y )

Correlation and Regression Statistical distinction based on nature of the variables –In correlation, both X and Y are random –In regression, X is fixed and Y is random Practical distinction based on interest of researcher –With correlation, the researcher asks: What is the strength (and direction) of the linear relationship between X and Y –With regression, the research asks the above and/or: How do I predict Y given X?

Goodness of Fit The regression equation does not reveal how well your data fit your model –e.g., in the below, both sets of data produce the same regression equation

Standard Error of Estimate The standard residual Why df = n - 2? –To determine regression equation (and thus the residuals), we need to estimate two population parameters Slope and intercept OR Mean of X and mean of Y –A regression with n = 2 has no df ^

Coefficicent of Determination (r 2 )

Partitioning Sums of Squares

Testing the Model # predictors n minus # model parameters n minus (1 + # predictors)

Online Applets Explaining variance – /tiein/johnson/reg.htmhttp:// /tiein/johnson/reg.htm Leverage – ssion.htmlhttp:// ssion.html Distribution of slopes/intercepts – t003.htmlhttp://lstat.kuleuven.be/java/version2.0/Apple t003.html