Prediction II Assumptions and Interpretive Aspects.

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

Kin 304 Regression Linear Regression Least Sum of Squares
Forecasting Using the Simple Linear Regression Model and Correlation
Chi Square Tests Chapter 17.
Monday, November 9 Correlation and Linear Regression.
Statistics Measures of Regression and Prediction Intervals.
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Statistical Tests Karen H. Hagglund, M.S.
Data Analysis Statistics. Inferential statistics.
Correlation. Introduction Two meanings of correlation –Research design –Statistical Relationship –Scatterplots.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Chapter 10 Simple Regression.
PPA 501 – Analytical Methods in Administration Lecture 8 – Linear Regression and Correlation.
PPA 415 – Research Methods in Public Administration
Lecture 11 PY 427 Statistics 1 Fall 2006 Kin Ching Kong, Ph.D
REGRESSION What is Regression? What is the Regression Equation? What is the Least-Squares Solution? How is Regression Based on Correlation? What are the.
PSY 1950 Correlation November 5, Definition Correlation quantifies the strength and direction of a linear relationship between two variables.
Chapter Eighteen MEASURES OF ASSOCIATION
Chapter Topics Types of Regression Models
Data Analysis Statistics. Inferential statistics.
REGRESSION Predict future scores on Y based on measured scores on X Predictions are based on a correlation from a sample where both X and Y were measured.
Measures of Association Deepak Khazanchi Chapter 18.
Multiple Regression Research Methods and Statistics.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Simple Linear Regression Analysis
Prediction. Prediction is a scientific guess of unknown by using the known data. Statistical prediction is based on correlation. If the correlation among.
Correlation and Linear Regression
McGraw-Hill/Irwin Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 13 Linear Regression and Correlation.
Leedy and Ormrod Ch. 11 Gray Ch. 14
Chapter 12 Correlation and Regression Part III: Additional Hypothesis Tests Renee R. Ha, Ph.D. James C. Ha, Ph.D Integrative Statistics for the Social.
Introduction to Linear Regression and Correlation Analysis
Equations in Simple Regression Analysis. The Variance.
Chapter 13 Statistics © 2008 Pearson Addison-Wesley. All rights reserved.
Chapter 6 & 7 Linear Regression & Correlation
© 2008 Pearson Addison-Wesley. All rights reserved Chapter 1 Section 13-6 Regression and Correlation.
© 2014 by Pearson Higher Education, Inc Upper Saddle River, New Jersey All Rights Reserved HLTH 300 Biostatistics for Public Health Practice, Raul.
Hypothesis of Association: Correlation
Correlation and Prediction Error The amount of prediction error is associated with the strength of the correlation between X and Y.
Examining Relationships in Quantitative Research
1 Review of ANOVA & Inferences About The Pearson Correlation Coefficient Heibatollah Baghi, and Mastee Badii.
1 Inferences About The Pearson Correlation Coefficient.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
Chapter 16 Data Analysis: Testing for Associations.
Psychology 820 Correlation Regression & Prediction.
Examining Relationships in Quantitative Research
CHAPTER 5 CORRELATION & LINEAR REGRESSION. GOAL : Understand and interpret the terms dependent variable and independent variable. Draw a scatter diagram.
© Buddy Freeman, 2015 Let X and Y be two normally distributed random variables satisfying the equality of variance assumption both ways. For clarity let.
Introducing Communication Research 2e © 2014 SAGE Publications Chapter Seven Generalizing From Research Results: Inferential Statistics.
Correlation They go together like salt and pepper… like oil and vinegar… like bread and butter… etc.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Simple Linear Regression Analysis Chapter 13.
Linear Regression and Correlation Chapter GOALS 1. Understand and interpret the terms dependent and independent variable. 2. Calculate and interpret.
© 2011 Cengage Learning. All Rights Reserved. May not be copied, scanned, or duplicated, in whole or in part, except for use as permitted in a license.
Multiple Regression Analysis Regression analysis with two or more independent variables. Leads to an improvement.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 7: Regression.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Linear Regression and Correlation Chapter 13.
Construction Engineering 221 Probability and Statistics.
LESSON 4.1. MULTIPLE LINEAR REGRESSION 1 Design and Data Analysis in Psychology II Salvador Chacón Moscoso Susana Sanduvete Chaves.
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
McGraw-Hill/Irwin © 2003 The McGraw-Hill Companies, Inc.,All Rights Reserved. Part Four ANALYSIS AND PRESENTATION OF DATA.
Correlation and Linear Regression Chapter 13 McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
REGRESSION G&W p
Chapter 15 Linear Regression
BIVARIATE REGRESSION AND CORRELATION
Multiple Regression A curvilinear relationship between one variable and the values of two or more other independent variables. Y = intercept + (slope1.
Section 6.2 Prediction.
Presentation transcript:

Prediction II Assumptions and Interpretive Aspects

Assumptions of Regression Normal Distribution – Both variables should be normally distributed – For non-normal distributions we use non-parametric tests Continuous Variables – Variables must be measured with a interval or ratio scale – Non-parametric tests are better for the scores collected with a nominal and ordinal scales Linearity – The relation between two variables should be linear Homoscedasticity – The variability of of actual Y values about Y I must be the same for all values of X.

Linearity In unlinear distributions, r is lower than its real value – So, prediction is less successful Some characteristics in nature are curvilinearly related. For such variables, we need to use some advanced tecniques For instance, the relationship between anxiety and success is curvilinear When anxiety is low, success is low (motivation is low) When anxiety is at its medium, success is high (motivation is high and anxiety does not have a derograting effect) When anxiety is high, success is low (the organism is shocked)

Homoscedasticity

Interpretive Aspects Factors Influencing r Range of Talent – When Y, X or both are restricted the r is lower than its real value – Because, r is a byproduct of both S 2 YX and S 2 Y That is S 2 YX / S 2 Y in formula B If we restrict the variance of Y, for instance, standart error of prediction would stay same. So, the r would get lower – See figure 11.1 on page 195 – This is what we called ceiling and floor effect

Interpretive Aspects Factors Influencing r Range of Talent Item per minute WorkerPayment $ApprerentReal A1288 B1311 C1412 D1512 E1613 F G1815 I1915 J K L M N r=0,840,98

Interpretive Aspects Factors Influencing r Heterogeneity of Samples – When samples are pooled, the correlation for aggregated data depends on where the sample values lie relative to one another in both the X and Y dimensions Let’s say professor Aktan and Göktürk prepared final exams for two courses: Statistics and Int. Resch. Methd.

Interpretive Aspects Factors Influencing r Heterogeneity of Samples – Students always gets 20 points higher in Göktürk’s exams StatisticsResearch Göktürk r=0,95 Aktan r=0,95 r=0,98

Interpretive Aspects Factors Influencing r Heterogeneity of Samples – Aktan insist on giving his own Statistics exam StatisticsResearch Göktürk r=0,95 Aktan r=0,95 r=0,58

Interpretive Aspects Regression Equation β coefficient shows the slope of the regression line. – General equation of a straight line Y=bX + c – Regression of Y on X

Interpretive Aspects Regression Equation β coefficient shows the slope of the regression line. – To see that let’s use two z score distribution in which mean is 0 and SD is 1 – Now, Zx-mean and Zy-mean becomes 0. So, c=0 – Zsy/Zsx is equal to 1/1. So, B=(r1/1)Zx= rZx – As you can see, beta is equal to r in z distributions

Interpretive Aspects Regression Equation Now, let’s say we calculated r between statistics and research scores for students of Çağ, ODTÜ and Mersin University – For Çağ University r=.82 – For Mersin University r=.62 – For ODTÜ r=.35 StatisticsResearch Çağ -3-2, ,64 -0, ,82 21,64 32,46 Mersin -3-1, ,24 -0, ,62 21,24 31,86 ODTÜ -3-1, ,7 -0, ,35 20,7 31,05

Interpretive Aspects Proportion of Variance in Y Associated with Variance in X Correlation coefficient has a special meaning – The squared correlation coefficient is equal to the proportion of variance in Y which is explained by the variance in X That is explained variance – r 2 = proportion of explained variance – 1- r 2 = proportion of unexplained variance – Let’s say correlation between depression and GPA is.67 So, change in depression explains 45% of change in GPA – r=.67, so r 2 =.45

Interpretive Aspects Proportion of Variance in Y Associated with Variance in X We can see the meaning of this in the Figure below