Linear Regression (C7-9 BVD). * Explanatory variable goes on x-axis * Response variable goes on y-axis * Don’t forget labels and scale * Statplot 1 st.

Slides:



Advertisements
Similar presentations
Scatterplots and Correlation
Advertisements

Residuals.
Section 10-3 Regression.
Chapter 4 The Relation between Two Variables
Lesson Diagnostics on the Least- Squares Regression Line.
Chapter 3 Bivariate Data
2nd Day: Bear Example Length (in) Weight (lb)
Scatter Diagrams and Linear Correlation
CHAPTER 3 Describing Relationships
C HAPTER 2 S CATTER PLOTS, C ORRELATION, L INEAR R EGRESSION, I NFERENCES FOR R EGRESSION By: Tasha Carr, Lyndsay Gentile, Darya Rosikhina, Stacey Zarko.
Haroon Alam, Mitchell Sanders, Chuck McAllister- Ashley, and Arjun Patel.
C HAPTER 3: E XAMINING R ELATIONSHIPS. S ECTION 3.3: L EAST -S QUARES R EGRESSION Correlation measures the strength and direction of the linear relationship.
Linear Regression Analysis
Correlation & Regression
Descriptive Methods in Regression and Correlation
Linear Regression.
Relationship of two variables
Correlation with a Non - Linear Emphasis Day 2.  Correlation measures the strength of the linear association between 2 quantitative variables.  Before.
Chapter 3: Examining relationships between Data
Correlation Correlation measures the strength of the LINEAR relationship between 2 quantitative variables. Labeled as r Takes on the values -1 < r < 1.
Lesson Least-Squares Regression. Knowledge Objectives Explain what is meant by a regression line. Explain what is meant by extrapolation. Explain.
Notes Bivariate Data Chapters Bivariate Data Explores relationships between two quantitative variables.
Notes Bivariate Data Chapters Bivariate Data Explores relationships between two quantitative variables.
Chapters 8 & 9 Linear Regression & Regression Wisdom.
Objective: Understanding and using linear regression Answer the following questions: (c) If one house is larger in size than another, do you think it affects.
Describing Bivariate Relationships Chapter 3 Summary YMS AP Stats Chapter 3 Summary YMS AP Stats.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 3 Describing Relationships 3.2 Least-Squares.
WARM-UP Do the work on the slip of paper (handout)
Examining Bivariate Data Unit 3 – Statistics. Some Vocabulary Response aka Dependent Variable –Measures an outcome of a study Explanatory aka Independent.
S CATTERPLOTS Correlation, Least-Squares Regression, Residuals Get That Program ANSCOMBE CRICKETS GESSEL.
Chapter 3-Examining Relationships Scatterplots and Correlation Least-squares Regression.
3.2 Least-Squares Regression Objectives SWBAT: INTERPRET the slope and y intercept of a least-squares regression line. USE the least-squares regression.
AP Statistics HW: p. 165 #42, 44, 45 Obj: to understand the meaning of r 2 and to use residual plots Do Now: On your calculator select: 2 ND ; 0; DIAGNOSTIC.
Linear Regression Day 1 – (pg )
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 3 Association: Contingency, Correlation, and Regression Section 3.3 Predicting the Outcome.
^ y = a + bx Stats Chapter 5 - Least Squares Regression
CHAPTER 3 Describing Relationships
Chapter 8 Linear Regression. Fat Versus Protein: An Example 30 items on the Burger King menu:
Response Variable: measures the outcome of a study (aka Dependent Variable) Explanatory Variable: helps explain or influences the change in the response.
Chapters 8 Linear Regression. Correlation and Regression Correlation = linear relationship between two variables. Summarize relationship with line. Called.
Describing Relationships. Least-Squares Regression  A method for finding a line that summarizes the relationship between two variables Only in a specific.
Least Squares Regression Textbook section 3.2. Regression LIne A regression line describes how the response variable (y) changes as an explanatory variable.
Part II Exploring Relationships Between Variables.
1. Analyzing patterns in scatterplots 2. Correlation and linearity 3. Least-squares regression line 4. Residual plots, outliers, and influential points.
CHAPTER 3 Describing Relationships
Ch. 10 – Linear Regression (Day 2)
Unit 4 LSRL.
CHAPTER 3 Describing Relationships
Section 3.2: Least Squares Regression
Chapter 5 LSRL.
Chapter 4 Correlation.
Suppose the maximum number of hours of study among students in your sample is 6. If you used the equation to predict the test score of a student who studied.
Regression and Residual Plots
Day 13 Agenda: DG minutes.
Chapter 3: Describing Relationships
CHAPTER 3 Describing Relationships
^ y = a + bx Stats Chapter 5 - Least Squares Regression
CHAPTER 3 Describing Relationships
Least-Squares Regression
Least-Squares Regression
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
CHAPTER 3 Describing Relationships
Algebra Review The equation of a straight line y = mx + b
Chapters Important Concepts and Terms
CHAPTER 3 Describing Relationships
Presentation transcript:

Linear Regression (C7-9 BVD)

* Explanatory variable goes on x-axis * Response variable goes on y-axis * Don’t forget labels and scale * Statplot 1 st option, specify lists, Zoom 9 * Direction: Positive slope or negative slope * Unusual points – outliers, influential points * Shape – straight or curves * Scatter – weak, moderate, strong

* Correlation measures the strength of linear association between two variables. r =( Σz x z y )/(n-1) * r is always between -1 and 1, inclusive. Negative r means a negative direction, positive r means positive direction. If r is close to 0, that indicates a weaker association, and near 1 or -1 indicates a stronger association. If r is exactly 1 or -1, that means the data are exactly linear. * Correlation is affected a lot by outliers: always look at the scatterplot and residual plot to make sure a linear model makes sense!

* Use a “hat” over y variable to indicate it is a model’s prediction vs. an actual data value. * Use words for y and x variables * y = mx + b but statisticians may use b 1 for slope and b 0 for y-intercept, and the calculator may use b for slope and a for y-intercept * To find a predicted y, plug x into equation and find y! * Extrapolation, or using your model to make predictions for x’s far from the x’s used to make the model, is likely to lead to inaccurate or worthless predictions.

* Residual = actual y value for a given x minus y value model predicts for a given x * Positive residual means model underestimated, negative residual means model overestimated * Least Squares Regression is a procedure to get line of best fit that minimizes the sum of the squares of residuals. * LinReg L 1,L 2,Y 1 in calculator (make sure diagnostics are on to get r and r 2 )

* When doing Linear Regression, always check the scatterplot (DUSS) to make sure shape is basically straight and ALSO check residuals plot for any patterns. * RESID list is automatically generated by calculator whenever you do LINREG. It is under 2 nd Stat Edit. Be careful, it overwrites itself if you do a new LINREG. Graph by changing Y-list in Statplot to RESID. * You can also store the RESID list in another list like L 3 for easy viewing of the residuals. * If you find standard deviation of residuals (1-var stats would work), it represents the approximate size of a “typical” prediction error (how far off the model will typically be for a given x).

* If you have stats (mean, standard deviation of x and y) instead of data, use these formulas to find line of best fit: * b 1 = rS y /S x * b 0 = mean y – b 1 (mean x) * If you want to switch explanatory and response, you can use these equations with reversed x and y means and standard deviations to find the new model.

* ___% of the variation in y (context) is accounted for by the regression line of y on x (context).

* Outliers are points far from the other points. * Outliers that are far away in the y-direction but not x-direction have large residuals and may artificially raise r. * Outliers that are far away in the x-direction pull the regression line toward them and may artificially lower r. These are called influential points. * Consider removing influential points before regression and considering them separately. * Also consider alternative models if appropriate. * Do not assume large r means a linear model is THE best model for the data or that it implies a cause/effect relationship between the variables.