Psyc 235: Introduction to Statistics DON’T FORGET TO SIGN IN FOR CREDIT!

Slides:



Advertisements
Similar presentations
Chapter 12 Inference for Linear Regression
Advertisements

6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives Copyright © 2004 John Wiley & Sons, Inc. Bivariate Correlation and Regression CHAPTER Thirteen.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Correlation and Regression
Sociology 601 Class 17: October 28, 2009 Review (linear regression) –new terms and concepts –assumptions –reading regression computer outputs Correlation.
Overview Correlation Regression -Definition
Scatter Diagrams and Linear Correlation
© The McGraw-Hill Companies, Inc., 2000 CorrelationandRegression Further Mathematics - CORE.
Chapter 12 Simple Regression
The Simple Regression Model
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 13 Introduction to Linear Regression and Correlation Analysis.
Chapter 3 Summarizing Descriptive Relationships ©.
Chapter Topics Types of Regression Models
Linear Regression and Correlation Analysis
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Introduction to Probability and Statistics Linear Regression and Correlation.
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Correlation 1. Correlation - degree to which variables are associated or covary. (Changes in the value of one tends to be associated with changes in the.
Correlation and Regression Analysis
Simple Linear Regression and Correlation
Relationships Among Variables
Correlation & Regression Math 137 Fresno State Burger.
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Correlation & Regression
Correlation and Linear Regression
Correlation and Regression A BRIEF overview Correlation Coefficients l Continuous IV & DV l or dichotomous variables (code as 0-1) n mean interpreted.
Chapter 8: Bivariate Regression and Correlation
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 11 Regression.
Introduction to Linear Regression and Correlation Analysis
Linear Regression and Correlation
Correlation and Linear Regression
Psyc 235: Introduction to Statistics DON’T FORGET TO SIGN IN FOR CREDIT!
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
● Final exam Wednesday, 6/10, 11:30-2:30. ● Bring your own blue books ● Closed book. Calculators and 2-page cheat sheet allowed. No cell phone/computer.
Simple Linear Regression One reason for assessing correlation is to identify a variable that could be used to predict another variable If that is your.
© The McGraw-Hill Companies, Inc., Chapter 11 Correlation and Regression.
Introduction to Linear Regression
Chapter 10 Correlation and Regression
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Topic 10 - Linear Regression Least squares principle - pages 301 – – 309 Hypothesis tests/confidence intervals/prediction intervals for regression.
Basic Concepts of Correlation. Definition A correlation exists between two variables when the values of one are somehow associated with the values of.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Last Time:. 2/3 2/3 of all Type A respondents had measurements between 55 and 69.
Chapter Thirteen Copyright © 2006 John Wiley & Sons, Inc. Bivariate Correlation and Regression.
LECTURE 9 Tuesday, 24 FEBRUARY STA291 Fall Administrative 4.2 Measures of Variation (Empirical Rule) 4.4 Measures of Linear Relationship Suggested.
Scatter Diagrams scatter plot scatter diagram A scatter plot is a graph that may be used to represent the relationship between two variables. Also referred.
Chapter 14: Inference for Regression. A brief review of chapter 4... (Regression Analysis: Exploring Association BetweenVariables )  Bi-variate data.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
STATISTICS 12.0 Correlation and Linear Regression “Correlation and Linear Regression -”Causal Forecasting Method.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
BUSINESS MATHEMATICS & STATISTICS. Module 6 Correlation ( Lecture 28-29) Line Fitting ( Lectures 30-31) Time Series and Exponential Smoothing ( Lectures.
Chapter 13 Linear Regression and Correlation. Our Objectives  Draw a scatter diagram.  Understand and interpret the terms dependent and independent.
Linear Regression Essentials Line Basics y = mx + b vs. Definitions
Inference about the slope parameter and correlation
Data Analysis Module: Correlation and Regression
Correlation and Regression Basics
Correlation and Regression Basics
Correlation and Simple Linear Regression
Regression Models - Introduction
Lecture Notes The Relation between Two Variables Q Q
Correlation and Simple Linear Regression
Correlation and Regression
Simple Linear Regression
CHAPTER 12 More About Regression
Simple Linear Regression and Correlation
Presentation transcript:

Psyc 235: Introduction to Statistics DON’T FORGET TO SIGN IN FOR CREDIT!

(Not) Relationships among Variables Descriptive stats (e.g., mean, median, mode, standard deviation)  describe a sample of data z-test &/or t-test for a single population parameter (e.g., mean)  infer the true value of a single variable  ex: mean # of random digits that people can memorize

Relationships among Variables But relationships among more than one variable are the crucial feature of almost all scientific research. Examples:  How does the perception of a stimulus vary with the physical intensity of that stimulus?  How does the attitude towards the President vary with the socio-economic properties of the survey respondent?  How does the performance on a mental task vary with age?

Relationships among Variables More Examples:  How does depression vary with number of traumatic experiences?  How does undergraduate drinking vary with performance in quantitative courses?  How does memory performance vary with attention span?  etc... We’ve already learned a few ways to analyze relationships among 2 variables.

Relationships among Two Variables: Chi-Square Chi-Square test of independence (2-way contingency table) compare observed cell frequencies to the cell frequencies you’d expect if the two variables are independent. ex: X=geographical region: West coast, Midwest, East coast Y=favorite color: red, blue, green Note: both variables are categorical

Relationships among Two Variables: Chi-Square Observed frequencies: Expected frequencies:

Relationships among Two Variables: Chi-Square Observed frequencies: Expected frequencies: ≈17.97 if this exceeds critical value, reject H 0 that the 2 variables are independent (unrelated)

Relationships among Two Variables: z, t tests z-test &/or t-test for difference of population means compare values of one variable (Y) for 2 different levels/groups of another variable (X) ex: X=age: young people vs. old people Y=# random digits can memorize Q: Is the mean # digits the same for the 2 age groups?

Relationships among Two Variables: ANOVA ANOVA compare values of one variable (Y) for 3+ different levels/groups of another variable (X) ex: X=age: young people, middle-aged, old people Y=# random digits can memorize Q: Is the mean # digits the same for all 3 age groups?

Relationships among Two Variables: z, t & ANOVA NOTE: for z/t tests for differences, and for ANOVA, there are a small number of possible values for one of the variables (X) z, t ANOVA

Relationships among Two Variables: z, t & ANOVA NOTE: for z/t tests for differences, and for ANOVA, there are a small number of possible values for one of the variables (X) z, t ANOVA

Relationships among Two Variables: many values of X? What about when there are many possible values of BOTH variables? Maybe they’re even continuous (rather than discrete)? Correlation, and Simple Linear Regression will be used to analyze relationship among two such variables (scatter plot)

Correlation: Scatter Plots Does it look like there is a relationship?

Correlation measures the direction and strength of a linear relationship between two variables that is, it answers in a general way the question: “as the values of one variable change, how do the corresponding values of the other variable change?”

Linear Relationship linear relationship:  y=a +bx (straight line) Linear Not (strictly) Linear

Correlation Coefficient: r sign: direction of relationship magnitude (number): strength of relationship -1 ≤ r ≤ 1 r=0 is no linear relationship r=-1 is “perfect” negative correlation r=1 is “perfect” positive correlation Notes:  Symmetric Measure (You can exchange X and Y and get the same value)  Measures linear relationship only

Correlation Coefficient: r Formula: alt. formula (ALEKS): standardized values

Correlation: Examples Population: undergraduates

Correlation: Examples Population: undergraduates

Correlation: Examples Population: undergraduates

Correlation: Examples Others?

Correlation: Interpretation Correlation ≠ Causation!

Correlation: Interpretation Correlation ≠ Causation! When 2 variables are correlated, the causality may be:  X --> Y  X <-- Y  Z --> X&Y (“lurking” third variable)  or a combination of the above Examples: ice cream & murder, violence & video games, SAT verbal & math, booze & GPA Inferring causation requires consideration of: how data gathered (e.g., experiment vs. observation), other relevant knowledge, logic...

Simple Linear Regression PREDICTING one variable (Y) from another (X)  No longer symmetric like Correlation One variable is used to “explain” another variable X Variable Independent Variable Explaining Variable Exogenous Variable Predictor Variable Y Variable Dependent Variable Response Variable Endogenous Variable Criterion Variable

Simple Linear Regression idea: find a line (linear function) that best fits the scattered data points this will let us characterize the relationship between X & Y, and predict new values of Y for a given X value.

(0,a) b Intercept Slope bX+a X Reminder: (Simple) Linear Function Y=a+bX We are interested in this to model the relationship between an independent variable X and a dependent variable Y Y

1 X Y Simple Linear Regression all data points would fall right on the line

X Y A guess at the location of the regression line

X Y Another guess at the location of the regression line (same slope, different intercept)

X Y Initial guess at the location of the regression line

X Y Another guess at the location of the regression line (same intercept, different slope)

X Y Initial guess at the location of the regression line

X Y Another guess at the location of the regression line (different intercept and slope, same “center”)

X Y We will end up being reasonably confident that the true regression line is somewhere in the indicated region.

X Y Estimated Regression Line errors/residuals

X Y Estimated Regression Line

X Y Wrong Picture! Error Terms have to be drawn vertically

X Y Estimated Regression Line =“y hat”: predicted value of Y for X i

Estimating the Regression Line Idea: find the formula for the line that minimizes the squared errors  error: distance between actual data point and predicted value Y=a+bX Y=b 0 +b 1 x b 1 =slope of regression line b 0 =Y intercept of regression line

ALEKS: Y=b 0 +b 1 X b 1 (slope) b 0 (Y intercept) using correlation coefficient