URBP 204A QUANTITATIVE METHODS I Statistical Analysis Lecture IV Gregory Newmark San Jose State University (This lecture is based on Chapters 5,12,13,

Slides:



Advertisements
Similar presentations
Bivariate Analyses.
Advertisements

Bivariate Analysis Cross-tabulation and chi-square.
Correlation Chapter 9.
 What is chi-square  CHIDIST  Non-parameteric statistics 2.
S519: Evaluation of Information Systems
PSY 307 – Statistics for the Behavioral Sciences
PSY 340 Statistics for the Social Sciences Chi-Squared Test of Independence Statistics for the Social Sciences Psychology 340 Spring 2010.
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Chapter 5: Correlation Coefficients
PY 427 Statistics 1Fall 2006 Kin Ching Kong, Ph.D Lecture 12 Chicago School of Professional Psychology.
CORRELATION COEFFICIENTS What Does a Correlation Coefficient Indicate? What is a Scatterplot? Correlation Coefficients What Could a Low r mean? What is.
Social Research Methods
S519: Evaluation of Information Systems Social Statistics Ch5: Correlation.
PSY 307 – Statistics for the Behavioral Sciences Chapter 19 – Chi-Square Test for Qualitative Data Chapter 21 – Deciding Which Test to Use.
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Inferential Statistics
Leedy and Ormrod Ch. 11 Gray Ch. 14
Analyzing Data: Bivariate Relationships Chapter 7.
Understanding Research Results
Statistics for the Social Sciences Psychology 340 Fall 2013 Thursday, November 21 Review for Exam #4.
AM Recitation 2/10/11.
Estimation and Hypothesis Testing Faculty of Information Technology King Mongkut’s University of Technology North Bangkok 1.
Statistical Analysis I have all this data. Now what does it mean?
Inferential Statistics: SPSS
Selecting the Correct Statistical Test
LEARNING PROGRAMME Hypothesis testing Intermediate Training in Quantitative Analysis Bangkok November 2007.
Part IV Significantly Different: Using Inferential Statistics
Bivariate Relationships Analyzing two variables at a time, usually the Independent & Dependent Variables Like one variable at a time, this can be done.
Research Methods for Counselors COUN 597 University of Saint Joseph Class # 9 Copyright © 2014 by R. Halstead. All rights reserved.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Which Test Do I Use? Statistics for Two Group Experiments The Chi Square Test The t Test Analyzing Multiple Groups and Factorial Experiments Analysis of.
Statistics 11 Correlations Definitions: A correlation is measure of association between two quantitative variables with respect to a single individual.
Bivariate Correlation Lesson 11. Measuring Relationships n Correlation l degree relationship b/n 2 variables l linear predictive relationship n Covariance.
Statistics in Applied Science and Technology Chapter 13, Correlation and Regression Part I, Correlation (Measure of Association)
Statistical Analysis Topic – Math skills requirements.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Chapter 16 The Chi-Square Statistic
Hypothesis testing Intermediate Food Security Analysis Training Rome, July 2010.
Psych 230 Psychological Measurement and Statistics Pedro Wolf September 23, 2009.
Correlation Analysis. Correlation Analysis: Introduction Management questions frequently revolve around the study of relationships between two or more.
Chi-Square X 2. Parking lot exercise Graph the distribution of car values for each parking lot Fill in the frequency and percentage tables.
CLASS 5 Normal Distribution Hypothesis Testing Test between means of related groups Test between means of different groups Analysis of Variance (ANOVA)
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
URBP 204A QUANTITATIVE METHODS I Statistical Analysis Lecture III Gregory Newmark San Jose State University (This lecture accords with Chapters 9,10, &
CHI SQUARE TESTS.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
Chapter Outline Goodness of Fit test Test of Independence.
Chapter Eight: Using Statistics to Answer Questions.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Chi Square & Correlation
Chapter 7 Calculation of Pearson Coefficient of Correlation, r and testing its significance.
Copyright c 2001 The McGraw-Hill Companies, Inc.1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent variable.
Bullied as a child? Are you tall or short? 6’ 4” 5’ 10” 4’ 2’ 4”
Chapter 14 – 1 Chi-Square Chi-Square as a Statistical Test Statistical Independence Hypothesis Testing with Chi-Square The Assumptions Stating the Research.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 11 Testing for Differences Differences betweens groups or categories of the independent.
Chapter 13 Understanding research results: statistical inference.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
S519: Evaluation of Information Systems Social Statistics Inferential Statistics Chapter 15: Chi-square.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Interpretation of Common Statistical Tests Mary Burke, PhD, RN, CNE.
Appendix I A Refresher on some Statistical Terms and Tests.
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE
Hypothesis Testing Review
Last class Tutorial 1 Census Overview
URBP 204 A Class 6 CLASS 5 Test between means of related groups
UNDERSTANDING RESEARCH RESULTS: STATISTICAL INFERENCE
COMPARING VARIABLES OF ORDINAL OR DICHOTOMOUS SCALES: SPEARMAN RANK- ORDER, POINT-BISERIAL, AND BISERIAL CORRELATIONS.
CLASS 6 CLASS 7 Tutorial 2 (EXCEL version)
Presentation transcript:

URBP 204A QUANTITATIVE METHODS I Statistical Analysis Lecture IV Gregory Newmark San Jose State University (This lecture is based on Chapters 5,12,13, & 15 of Neil Salkind’s Statistics for People who (Think They) Hate Statistics, 2 nd Edition which is also the source of many of the offered examples. All cartoons are from CAUSEweb.org by J.B. Landers.)

More Statistical Tests Factorial Analysis of Variance (ANOVA) – Tests between means of more than two groups for two or more factors (independent variables) Correlation Coefficient – Tests the association between two variables One Sample Chi-Square (χ 2 ) – Tests if an observed distribution of frequencies for one factor is what one would expect by chance Two Factor Chi-Square (χ 2 ) – Tests if an observed distribution of frequencies for two factors is what one would expect by chance

Factorial ANOVA Compares observations of a single variable among two or more groups which incorporate two or more factors. Examples: – Reading Skills School (Elementary, Middle, High) Academic Philosophy (Montessori, Waldorf) – Environmental Knowledge Commute Mode (Car, Bus, Walking) Age (Under 40, 40+) – Wealth Favorite Team (A’s, Giants, Dodger, Angels) Home Location (Oakland, SF, LA) – Weight Loss Gender (Male, Female) Exercise (Biking, Running)

Factorial ANOVA Two Types of Effects – Main Effects: differences within one factor – Interaction Effects: differences across factors Example: – Weight Loss Gender (Male, Female) Exercise (Biking, Running) – Main Effects: Does weight loss vary by exercise? Does weight loss vary by gender? – Interaction Effects: Does weight loss due to exercise vary by gender?

Factorial ANOVA Example: – “How is weight loss affected by exercise program and gender?” Steps: – State hypotheses Null : H 0 : µ Male = µ Female H 0 : µ Biking = µ Running H 0 : µ Male-Biking = µ Female-Biking = µ Male-Running = µ Female-Running Research : What would these three be?

Factorial ANOVA Steps (Continued): – Set significance level Level of risk of Type I Error = 5% Level of Significance (p) = 0.05 – Select statistical test Factorial ANOVA – Computation of obtained test statistic value Insert obtained data into appropriate formula (SPSS can expedite this step for us)

Factorial ANOVA Weight Loss Data Male-BikingMale-RunningFemale-BikingFemale-Running

Factorial ANOVA SPSS Outputs =p

Factorial ANOVA SPSS Outputs

Factorial ANOVA SPSS Outputs – Graph them!

Factorial ANOVA Steps (Continued) – Computation of obtained test statistic value ExerciseF = 2.444, p = Gender F = 1.908, p = InteractionF = 9.683, p = – Look up the critical F score df numerator = # of Factors – 1 df denominator = # of Observations – # of Groups What is the critical F score? – Comparison of obtained and critical values If obtained > critical reject the null hypothesis If obtained < critical stick with the null hypothesis

Factorial ANOVA Steps (Continued) – Therefore we reject the null hypothesis for the interaction effects. This means that while choice of exercise alone and gender alone make no difference to weight loss, in combination they do differentially affect weight loss. Men should run and women should bike, according to these data.

Correlation Coefficient Tests whether changes in two variables are related Examples – “Are property values positively related to distance from waste dumps?” – “Is age correlated with height for minors?” – “Are apartment rents negatively related to commute time?” – “Does someone’s height relate to income?” – “How related are hand size and height?”

Correlation Coefficient Are Tastiness and Ease correlated for fruit? Is there directionality?

Correlation Coefficient Numeric index that reflects the linear relationship between two variables (bivariate correlation) – “How does the value of one variable change when another variable changes?” – Each case has two data points: E.g. This study records each persons height and weight to see if they are correlated. – Ranges from -1.0 to +1.0 – Two types of possible correlations Change in the same direction : positive or direct correlation Change in opposite directions: negative or indirect correlation – Absolute value reflects strength of correlation Pearson Product-Moment Correlation – Both variables need to be ratio or interval

Correlation Coefficient Scatterplot

Correlation Coefficient Coefficient of Determination – Squaring the correlation coefficient (r 2 ) – The percentage of variance in one variable that is accounted for by the variance in another variable Example: GPA and Time Spent Studying – [r GPA and Study Time = 0.70]; [r 2 GPA and Study Time = 0.49] 49% of the variance in GPA can be explained by the variance in studying time GPA and studying time share 49% of the variance between themselves

Correlation Coefficient Example – “How related are hand size and height?” Steps – State hypotheses Null : H 0 : ρ Hand Size and Height = 0 Research: H 1 : r Hand Size and Height ≠ 0 – Non-directional – Set significance level Level of risk of Type I Error = 5% Level of Significance (p) = 0.05

Correlation Coefficient Steps (Continued) – Select statistical test Correlation Coefficient (it is the test statistic!) – Computation of obtained test statistic value Insert obtained data into appropriate formula

Correlation Coefficient Plot the data: n = 30

Correlation Coefficient Steps (Continued) – Computation of obtained test statistic value r Hand Size and Height = Correlations HeightHand Height Pearson Correlation ** Sig. (2-tailed).000 N30 Hand Pearson Correlation.736 ** 1 Sig. (2-tailed).000 N30 **. Correlation is significant at the 0.01 level (2-tailed).

Correlation Coefficient Steps (Continued) – Computation of critical test statistic value Value needed to reject null hypothesis Look up p = 0.05 in critical value table Consider degrees of freedom [df= n – 2] Consider number of tails (is there directionality?) r critical = ?

Correlation Coefficient What happens to the critical score when the number of cases (n) decreases? Why?

Correlation Coefficient Steps (Continued) – Comparison of obtained and critical values If obtained > critical reject the null hypothesis If obtained < critical stick with the null hypothesis r obtained = > r critical = – Therefore, we reject the null hypothesis and accept the research hypothesis that height and handbreadth are correlated. Is there a directionality to that correlation?

Correlation Coefficient Significance vs. Meaning – Rules of Thumb r = 0.8 to 1.0Very strong relationship r = 0.6 to 0.8Strong relationship r = 0.4 to 0.6Moderate relationship r = 0.2 to 0.4Weak relationship r = 0.0 to 0.2Weak or no relationship

Correlation Coefficient Does correlation express causation? Classic Example: – Ice Cream Eaten – Crimes Committed

Correlation Coefficient Correlation expresses association only

Chi-Square (χ 2 ) Non-Parametric Test – Does not rely on a given distribution Useful for small sample sizes – Enables consideration of data that comes as ordinal or nominal frequencies Number of children in different grades Percentage of people by state receiving social security

One Sample Chi-Square (χ 2 ) Tests whether an observed distribution of frequencies for one factor is likely to have occurred by chance Examples: – “Is this community evenly distributed among ethnic groups?” – “Are the 31 ice cream flavors at Baskin Robbins equally purchased?” – “Are commuting mode shares evenly spread out?” – “Did people report equal preferences for a school voucher policy?”

One Sample Chi-Square (χ 2 ) Examples: – “Did people report equal preferences for a school voucher policy?” – Data (90 People split into 3 Categories) For23 Maybe 17 Against50 – Always try to have at least 5 responses per category

One Sample Chi-Square (χ 2 ) Steps: – State hypotheses Null : H 0 : Proportion For = Proportion Maybe = Proportion Against Research : H 1 : Proportion For ≠ Proportion Maybe ≠ Proportion Against – Set significance level Level of risk of Type I Error = 5% Level of Significance (p) = 0.05 – Select statistical test Chi-Square (χ 2 )

One Sample Chi-Square (χ 2 ) Steps (Continued): – Computation of obtained test statistic value Insert obtained data into appropriate formula (SPSS can expedite this step for us)

One Sample Chi-Square (χ 2 ) Steps (Continued): – Computation of obtained test statistic value CategoryOE(O-E)(O-E) 2 (O-E) 2 /E For Against Maybe Total

One Sample Chi-Square (χ 2 ) Steps (Continued): – Computation of obtained test statistic value χ 2 obtained = – Computation of critical test statistic value Value needed to reject null hypothesis Look up p = 0.05 in χ 2 table Consider degrees of freedom [df= # of categories - 1] χ 2 critical = 5.99

One Sample Chi-Square (χ 2 ) Steps (Continued): – Computation of obtained test statistic value

One Sample Chi-Square (χ 2 ) Steps (Continued): – Comparison of obtained and critical values If obtained > critical reject the null hypothesis If obtained < critical stick with the null hypothesis χ 2 obtained = > χ 2 critical = 5.99 – Therefore, we can reject the null hypothesis and we thus conclude that distribution of preferences regarding the school voucher is not even.

Two Factor Chi-Square (χ 2 ) What if we want to see if gender effects the distribution of votes? How is this different from Factorial ANOVA?

Two Factor Chi-Square (χ 2 ) Steps: – State hypotheses Null : H 0 : P For*Male = P Maybe*Male = P Against *Male = P For*Female = P Maybe*Female = P Against *Female Research : H 1 : P For*Male ≠ P Maybe*Male ≠ P Against *Male ≠ P For*Female ≠ P Maybe*Female ≠ P Against *Female – Set significance level Level of risk of Type I Error = 5% Level of Significance (p) = 0.05 – Select statistical test Chi-Square (χ 2 )

Two Factor Chi-Square (χ 2 ) Steps (Continued): – Computation of obtained test statistic value Insert obtained data into appropriate formula Same as for One Factor Chi-Square

Two Factor Chi-Square (χ 2 ) How do we find the expected frequencies? – (Row Total * Column Total)/ Total Total – Expected Value [For*Male] = (23*44)/90 = 11.2

Two Factor Chi-Square (χ 2 ) Steps (Continued): – Computation of obtained test statistic value χ 2 obtained = 7.750

Two Factor Chi-Square (χ 2 ) Steps (Continued): – Computation of critical test statistic value Value needed to reject null hypothesis Look up p = 0.05 in χ 2 table Consider degrees of freedom df= (# of rows – 1) * (# of columns – 1) χ 2 critical = ?

Two Factor Chi-Square (χ 2 ) Steps (Continued): – Comparison of obtained and critical values If obtained > critical reject the null hypothesis If obtained < critical stick with the null hypothesis χ 2 obtained = > χ 2 critical = 5.99 – Therefore, we can reject the null hypothesis and we thus conclude that gender affects the distribution of preferences regarding the school vouchers.

Tutorial Time