10/22/20151 PUAF 610 TA Session 8. 10/22/20152 Recover from midterm.

Slides:



Advertisements
Similar presentations
Hypothesis Testing Steps in Hypothesis Testing:
Advertisements

Learning Objectives Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Learning Objectives 1 Copyright © 2002 South-Western/Thomson Learning Data Analysis: Bivariate Correlation and Regression CHAPTER sixteen.
Correlation and Regression
Describing Relationships Using Correlation and Regression
Correlation Chapter 9.
Chapter 15 (Ch. 13 in 2nd Can.) Association Between Variables Measured at the Interval-Ratio Level: Bivariate Correlation and Regression.
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Statistics Are Fun! Analysis of Variance
Correlation Patterns. Correlation Coefficient A statistical measure of the covariation or association between two variables. Are dollar sales.
The Simple Regression Model
SIMPLE LINEAR REGRESSION
Chapter Topics Types of Regression Models
Regression Chapter 10 Understandable Statistics Ninth Edition By Brase and Brase Prepared by Yixun Shi Bloomsburg University of Pennsylvania.
Inferences About Process Quality
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Chapter 7 Forecasting with Simple Regression
Linear Regression/Correlation
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Nonparametric or Distribution-free Tests
Correlation and Linear Regression
Leedy and Ormrod Ch. 11 Gray Ch. 14
Lecture 16 Correlation and Coefficient of Correlation
AM Recitation 2/10/11.
F-Test ( ANOVA ) & Two-Way ANOVA
The Chi-Square Distribution 1. The student will be able to  Perform a Goodness of Fit hypothesis test  Perform a Test of Independence hypothesis test.
Introduction to Linear Regression and Correlation Analysis
Correlation and Linear Regression
CENTRE FOR INNOVATION, RESEARCH AND COMPETENCE IN THE LEARNING ECONOMY Session 2: Basic techniques for innovation data analysis. Part I: Statistical inferences.
Correlation.
Sections 9-1 and 9-2 Overview Correlation. PAIRED DATA Is there a relationship? If so, what is the equation? Use that equation for prediction. In this.
Is there a relationship between the lengths of body parts ?
Chapter 13 Statistics © 2008 Pearson Addison-Wesley. All rights reserved.
Learning Objective Chapter 14 Correlation and Regression Analysis CHAPTER fourteen Correlation and Regression Analysis Copyright © 2000 by John Wiley &
EQT 272 PROBABILITY AND STATISTICS
1 1 Slide © 2005 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
© 2003 Prentice-Hall, Inc.Chap 13-1 Basic Business Statistics (9 th Edition) Chapter 13 Simple Linear Regression.
INTRODUCTORY LINEAR REGRESSION SIMPLE LINEAR REGRESSION - Curve fitting - Inferences about estimated parameter - Adequacy of the models - Linear.
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Hypothesis of Association: Correlation
ANOVA (Analysis of Variance) by Aziza Munir
Examining Relationships in Quantitative Research
Chapter 14 – 1 Chapter 14: Analysis of Variance Understanding Analysis of Variance The Structure of Hypothesis Testing with ANOVA Decomposition of SST.
1 Chapter 12 Simple Linear Regression. 2 Chapter Outline  Simple Linear Regression Model  Least Squares Method  Coefficient of Determination  Model.
Analysis of Variance 1 Dr. Mohammed Alahmed Ph.D. in BioStatistics (011)
CORRELATION: Correlation analysis Correlation analysis is used to measure the strength of association (linear relationship) between two quantitative variables.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 16 Data Analysis: Testing for Associations.
Chapter 13 Multiple Regression
Previous Lecture: Phylogenetics. Analysis of Variance This Lecture Judy Zhong Ph.D.
Lecture 10: Correlation and Regression Model.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
Chapter 9: Correlation and Regression Analysis. Correlation Correlation is a numerical way to measure the strength and direction of a linear association.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 11 Analyzing the Association Between Categorical Variables Section 11.2 Testing Categorical.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
Linear Regression and Correlation Chapter GOALS 1. Understand and interpret the terms dependent and independent variable. 2. Calculate and interpret.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
©The McGraw-Hill Companies, Inc. 2008McGraw-Hill/Irwin Linear Regression and Correlation Chapter 13.
REGRESSION AND CORRELATION SIMPLE LINEAR REGRESSION 10.2 SCATTER DIAGRAM 10.3 GRAPHICAL METHOD FOR DETERMINING REGRESSION 10.4 LEAST SQUARE METHOD.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Statistics for Managers using Microsoft Excel 3rd Edition
Correlation and Simple Linear Regression
Comparing Three or More Means
Correlation and Regression
Simple Linear Regression and Correlation
Warsaw Summer School 2017, OSU Study Abroad Program
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

10/22/20151 PUAF 610 TA Session 8

10/22/20152 Recover from midterm

10/22/20153 TODAY F - Distribution Analysis-of-variance Kruskal-Wallis test Correlation

10/22/20154 F - Distribution An F-distribution has two numbers of degrees of freedom : –Degrees of Freedom for the Numerator (dfn) –Degrees of Freedom for the Denominator (dfd) There is a different F distribution for each combination of the degrees of freedom of the numerator and denominator.

10/22/20155 F - Distribution

10/22/20156 F - Distribution Using F table –Select the significance level to be used –Determine the appropriate combination of degrees of freedom

10/22/20157 F - Distribution If the α = 0.10 level of significance is selected. There are 5 degrees of freedom in the numerator, and 7 degrees of freedom in the denominator. The F value from the table is This means that there is exactly 0.10 of the area under the F curve that lies to the right of F = 2.88

10/22/20158 Analysis-of-variance Introduction Question: Do the means of the quantitative variables depend on which group (given by categorical variable) the individual is in? –If categorical variable has only 2 values: 2- sample t-test –If categorical variable has 3 or more values: ANOVA

10/22/20159 Assumptions of ANOVA

10/22/ Logic of ANOVA Hypothesis F-statistic Compare CV or p Conclusion

10/22/ Analysis-of-variance ANOVA measures two sources of variation in the data and compares their relative sizes variation BETWEEN groups –the difference between its group mean and the overall mean variation WITHIN groups –the difference between that value and the mean of its group

10/22/ Analysis-of-variance The ANOVA F-statistic is a ratio of the “Between Group Variation” divided by the “Within Group Variation”: A large F indicates that there is more difference between groups than within groups. Treatment mean square Error mean square

10/22/ Analysis-of-variance

10/22/ Analysis-of-variance

10/22/201515

10/22/201516

10/22/201517

10/22/201518

10/22/ Example A manager wishes to determine whether the mean times required to complete a certain task differ for the three levels of employee training. He randomly selected 10 employees with each of the three levels of training (Beginner, Intermediate and Advanced). Do the data provide sufficient evidence to indicate that the mean times required to complete a certain task differ for at least two of the three levels of training?

10/22/ Example Level of Training N Advanced Intermediate Beginner

10/22/ Example Ho: The mean times required to complete a certain task do not differ the three levels of training. ( µB = µI = µA) Ha: The mean times required to complete a certain task differ for at least two of the three levels of training.

10/22/ Example SourcedfSSMSF Treatments Error Total

10/22/ Example

10/22/ Example Decision: Reject Ho. Conclusion: There is sufficient evidence to indicate that the mean times required to complete a certain task differ for at least two of the three levels of training.

Example Consider the following random samples from three different populations. Do the data provide sufficient evidence to indicate that the mean differ? 10/22/201525

Example The data below resulted from measuring the difference in resistance resulting from subjecting identical resistors to three different temperatures for a period of 24 hours. The sample size of each group was 5. In the language of Design of Experiments, we have an experiment in which each of three treatments was replicated 5 times. 10/22/201526

Example Level 1Level 2Level Is there a difference among the population means? 10/22/201527

10/22/ Kruskal-Wallis test The Kruskal-Wallis test is a nonparametric (distribution free) test, which is used to compare three or more groups of independent groups of sampled data. Kruskal-Wallis Test is used when assumptions of ANOVA are not met. –In ANOVA, we assume that distribution of each group should be normally distributed. –In Kruskal-Wallis Test, we do not assume the distribution. If normality assumptions are met, then the Kruskal-Wallis Test is not as powerful as ANOVA.

10/22/ Kruskal-Wallis test The hypotheses for the comparison of two independent groups are: –H0 : μ1 = μ2 = μ3 –Ha : Not all the means are equal.

10/22/ Kruskal-Wallis test The test statistic for the Kruskal-Wallis test is H. When sample sizes are small in each group (< 5) and the number of groups is less than 4, a tabled value for the Kruskal-Wallis should be compared to the H statistic to determine the significance level. Otherwise, a Chi-square with k-1 (the number of groups-1) degrees of freedom can be used to approximate the significance level for the test.

10/22/201531

10/22/201532

10/22/ Kruskal-Wallis test Do not worry. You just should know how to interpret the STATA output.

10/22/ When to use what?

10/22/ Correlation Find out the relationship of the two variables: if they are associated or not? Standard measure of the relationship is called correlation. This relationship could be –negative (with increased x y goes down) –positive (as x increases y also increases).

10/22/ Correlation Scatter plots provide a graphical way of interpreting the correlation. –Scatter plots are similar to line graphs in that they use horizontal and vertical axes to plot data points –The closer the data points come when plotted to making a straight line, the higher the correlation between the two variables, or the stronger the relationship.

10/22/ Correlation Scatter plots provide an approximation. –If the data points make a straight line going from the origin out to high x- and y- values, then the variables have a positive correlation. –If the line goes from a high- value on the y-axis down to a high-value on the x-axis, the variables have a negative correlation.

10/22/ Correlation Variables may not have a correlation at all. There is no pattern to where the data points lie. They do not seem to go in any particular direction.

10/22/ Correlation we use correlation coefficient to figure out the exact value for the relationship between two variables. –the strength and the direction of a linear relationship between two variables. Correlation coefficient (r) has a range of -1 to +1 The absolute value of the coefficient is a measure of magnitude.

10/22/ Correlation Positive correlation: –If x and y have a strong positive linear correlation, r is close to +1. –An r value of exactly +1 indicates a perfect positive fit. –As values for x increases, values for y also increase. Negative correlation: –If x and y have a strong negative linear correlation, r is close to -1. –An r value of exactly -1 indicates a perfect negative fit. –As values for x increase, values for y decrease. No correlation: –If there is no linear correlation or a weak linear correlation, r is close to 0.

10/22/201541

10/22/ Correlation Interpret the Correlation Coefficient ! Suppose that a research study that reported a correlation of r =.75 between compensation and number of years of education. What could you say about the relationship between these two variables?

10/22/ Correlation There is a positive relationship between level of education and compensation. This means that people with more education tend to earn higher salaries. Similarly, people with low levels of education tend to have correspondingly lower salaries.