Action Research Data Manipulation and Crosstabs

Slides:



Advertisements
Similar presentations
Chapter 18: The Chi-Square Statistic
Advertisements

2013/12/10.  The Kendall’s tau correlation is another non- parametric correlation coefficient  Let x 1, …, x n be a sample for random variable x and.
Finish Anova And then Chi- Square. Fcrit Table A-5: 4 pages of values Left-hand column: df denominator df for MSW = n-k where k is the number of groups.
Hypothesis Testing Steps in Hypothesis Testing:
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Bivariate Analysis Cross-tabulation and chi-square.
Hypothesis Testing IV Chi Square.
Chapter 13: The Chi-Square Test
Chapter 11 Contingency Table Analysis. Nonparametric Systems Another method of examining the relationship between independent (X) and dependant (Y) variables.
Analysis of frequency counts with Chi square
Copyright (c) Bani K. Mallick1 STAT 651 Lecture #17.
PSY 340 Statistics for the Social Sciences Chi-Squared Test of Independence Statistics for the Social Sciences Psychology 340 Spring 2010.
Test statistic: Group Comparison Jobayer Hossain Larry Holmes, Jr Research Statistics, Lecture 5 October 30,2008.
Chi-square Test of Independence
Student’s t statistic Use Test for equality of two means
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Measures of Central Tendency
Inferential Statistics
Chapter 11(1e), Ch. 10 (2/3e) Hypothesis Testing Using the Chi Square ( χ 2 ) Distribution.
AM Recitation 2/10/11.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Action Research Review
1 Psych 5500/6500 Chi-Square (Part Two) Test for Association Fall, 2008.
Week 10 Chapter 10 - Hypothesis Testing III : The Analysis of Variance
Which Test Do I Use? Statistics for Two Group Experiments The Chi Square Test The t Test Analyzing Multiple Groups and Factorial Experiments Analysis of.
Chi-Square as a Statistical Test Chi-square test: an inferential statistics technique designed to test for significant relationships between two variables.
Hypothesis of Association: Correlation
Chi-square (χ 2 ) Fenster Chi-Square Chi-Square χ 2 Chi-Square χ 2 Tests of Statistical Significance for Nominal Level Data (Note: can also be used for.
Basic concept Measures of central tendency Measures of central tendency Measures of dispersion & variability.
1 Chi-Square Heibatollah Baghi, and Mastee Badii.
Chapter 20 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 These tests can be used when all of the data from a study has been measured on.
Chapter 16 The Chi-Square Statistic
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Pearson Chi-Square Contingency Table Analysis.
TYPES OF STATISTICAL METHODS USED IN PSYCHOLOGY Statistics.
Education 793 Class Notes Presentation 10 Chi-Square Tests and One-Way ANOVA.
SW388R6 Data Analysis and Computers I Slide 1 Multiple Regression Key Points about Multiple Regression Sample Homework Problem Solving the Problem with.
© 2014 by Pearson Higher Education, Inc Upper Saddle River, New Jersey All Rights Reserved HLTH 300 Biostatistics for Public Health Practice, Raul.
Analysis of Qualitative Data Dr Azmi Mohd Tamil Dept of Community Health Universiti Kebangsaan Malaysia FK6163.
Nonparametric Statistics
Nonparametric Tests: Chi Square   Lesson 16. Parametric vs. Nonparametric Tests n Parametric hypothesis test about population parameter (  or  2.
CHI SQUARE TESTS.
Chi-square Test of Independence
Reasoning in Psychology Using Statistics Psychology
4 normal probability plots at once par(mfrow=c(2,2)) for(i in 1:4) { qqnorm(dataframe[,1] [dataframe[,2]==i],ylab=“Data quantiles”) title(paste(“yourchoice”,i,sep=“”))}
Nonparametric Tests of Significance Statistics for Political Science Levin and Fox Chapter Nine Part One.
12/23/2015Slide 1 The chi-square test of independence is one of the most frequently used hypothesis tests in the social sciences because it can be used.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Leftover Slides from Week Five. Steps in Hypothesis Testing Specify the research hypothesis and corresponding null hypothesis Compute the value of a test.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Week 6 Dr. Jenne Meyer.  Article review  Rules of variance  Keep unaccounted variance small (you want to be able to explain why the variance occurs)
Chapter 14 – 1 Chi-Square Chi-Square as a Statistical Test Statistical Independence Hypothesis Testing with Chi-Square The Assumptions Stating the Research.
Chapter 4 Selected Nonparemetric Techniques: PARAMETRIC VS. NONPARAMETRIC.
I. ANOVA revisited & reviewed
Chi-Square (Association between categorical variables)
BINARY LOGISTIC REGRESSION
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE
Qualitative data – tests of association
Hypothesis Testing Using the Chi Square (χ2) Distribution
The Chi-Square Distribution and Test for Independence
Different Scales, Different Measures of Association
Reasoning in Psychology Using Statistics
Chi Square (2) Dr. Richard Jackson
BIVARIATE ANALYSIS: Measures of Association Between Two Variables
Reasoning in Psychology Using Statistics
BIVARIATE ANALYSIS: Measures of Association Between Two Variables
Parametric versus Nonparametric (Chi-square)
Reasoning in Psychology Using Statistics
Chapter 18: The Chi-Square Statistic
Presentation transcript:

Action Research Data Manipulation and Crosstabs INFO 515 Glenn Booker INFO 515 Lecture #8

Parametric vs. Nonparametric Statistical tests fall into two broad categories – parametric & nonparametric Parametric methods Require data at higher levels of measurement - interval and/or ratio scales Are more mathematically powerful than nonparametric statistics But often require more assumptions about the data, such as having a normal distribution, or equal variances INFO 515 Lecture #8

Parametric vs. Nonparametric Nonparametric methods Use nominal or ordinal scale data Still allows us to test for a relationship, and its strength and direction (direction only if ordinal) Often has easier prerequisites for being tested (e.g. no distribution limits) Ratio or interval scale data may be recoded to become nominal or ordinal data, and hence be used with nonparametric tests INFO 515 Lecture #8

Significance and Association … are useful for inferring population values from samples (inferential statistics) Significance establishes whether chance can be ruled out as the most likely explanation of differences Association shows the nature, strength, and/or direction of the relationship between two (or among three or more) variables Need to show significance before association is meaningful INFO 515 Lecture #8

Common Tests of Significance We’ve been introduced to three common tests of significance: z test (large samples of ratio or interval data) t test (small samples of ratio or interval data) F test (ANOVA) Shortly we’ll explore a fourth one Pearson’s chi-square 2 (used for nominal or ordinal scale data) { is the Greek letter chi, pronounced ‘kye’, rhymes with ‘rye’} INFO 515 Lecture #8

Common Measures of Association Association measures often range in value from -1 to +1 (but not always!) Absence of association between variables generally means a result of 0 Examples Pearson’s r (for interval or ratio scale data) Yule’s Q (ordinal data in a 2x2 table) Gamma (ordinal – more than 2x2 table) {A “2x2” table has 2 rows and 2 columns of data.} INFO 515 Lecture #8

Common Measures of Association Notice these are all for nominal scale data Phi (, ‘fee’) (nominal data in a 2x2 table) Contingency Coefficient (nominal – table larger than 2x2) Cramer’s V (nominal - larger than 2x2) Lambda (l) - nominal data Eta () – nominal data INFO 515 Lecture #8

Significance and Association Tests of significance and measures of association are often used together But you can have statistical significance without having association INFO 515 Lecture #8

Significance and Association Examples Ratio data: You might use F to determine if there is a significant relationship, then use ‘r’ from a regression to measure its strength Ordinal data: You might run a chi-square to determine statistical significance in the frequencies of two variables, and then run a Yule’s Q to show the relationship between the variables INFO 515 Lecture #8

Crosstabs Brief digression to introduce crosstabs before discussing non-parametric methods Crosstabs are a table, often used to display data, sorted by two nominal or ordinal variables at once, to study the relationship between variables that have a small number of possible answers each Generally contains basic descriptive statistics, such as frequency counts and percentages INFO 515 Lecture #8

Crosstabs Used to check the distribution of data, and as a foundation for more complex tests Look for gaps or sparse data (little or no contribution to the data set) Rule of thumb - put independent variable in the columns and dependent variable in the rows INFO 515 Lecture #8

Percentages Can show both column and row percentages in crosstabs, rather than just frequency counts (or show both counts and percentages) Make sure percentages add to 100%! Raw frequency counts of variables don’t always provide an accurate picture Unequal numbers of subjects in groups (N) might make the numbers appear skewed INFO 515 Lecture #8

Crosstabs Example Open data set “GSS91 political.sav” Use Analyze / Descriptive Statistics / Crosstabs... Set the Row(s) as “region”, and the Column(s) as “relig” Note the default scope of an SPSS crosstab is to show frequency Counts, with row and column totals INFO 515 Lecture #8

Crosstabs Example INFO 515 Lecture #8

Crosstabs Example Repeat the same example with percentages selected under the “Cells…” button to get detailed data in each cell Percent within that region (Row) Percent within that religious preference (Column) Percent of total data set (divide by Total N) Gets a bit messy to show this much! INFO 515 Lecture #8

Crosstabs Example INFO 515 Lecture #8

Recoding An interval or ratio scaled variable, like age or salary, may have too many distinct values to use in a crosstab Recoding lets you combine values into a single new variable -- also called collapsing the codes Also helpful for creating histogram variables (e.g. ranges of age or income) INFO 515 Lecture #8

Recoding Example Use Transform / Recode / Into Different Variables… Move “age” from the dropdown list for the Numeric Variable Define the new Output Variable to have Name “agegroup” and Label “Age Group” Click “Change” button to use “agegroup” Click on “Old and New Values” button INFO 515 Lecture #8

Recoding Example For the Old Value, enter Range of 18 to 30 Assign this to a New Value of 1 Click on “Add” Repeat to define ages 31-50 as agegroup New Value 2, 51-75 as 3, and 76-200 as 4 Click “Continue” and now a new variable exists as defined INFO 515 Lecture #8

Recoding Example INFO 515 Lecture #8

Recoding Example Now generate a crosstab with “agegroup” as columns, and “region” as the rows INFO 515 Lecture #8

Second Recoding Example Prof. Yonker had a previous INFO515 class surveyed for their height (in inches) and desired salaries ($/yr) Rather than analyze ratio data with few frequencies larger than one, she recoded: Heights into: Dwarves for people below average height, and Giants for those above Desired salaries were recoded into Cheap and Expensive, again below and above average INFO 515 Lecture #8

Second Recoding Example The resulting crosstab was like this: INFO 515 Lecture #8

Pearson Chi Square Test The Chi Square test measures how much observed (actual) frequencies (fo) differ from “expected” frequencies (fe) Is a nonparametric test, a.k.a. the Goodness of Fit statistic Does not require assumptions about the shape of the population distribution Does not require variables be measured on an interval or ratio scale INFO 515 Lecture #8

Chi Square Concept Chi Square test is like the ANOVA test ANOVA proved whether there was a difference among several means – proved that the means are different from each other in some way Chi square is trying to prove whether the frequency distribution is different from a random one – is there a significant difference among frequencies? Allows us to test for a relationship (but not the strength or direction if there is one) INFO 515 Lecture #8

Chi Square Null Hypothesis Null hypothesis is that the frequencies in cells are independent of each other (there is no relationship among them) Each case is independent of every other case; that is, the value of the variable for one individual does not influence the value for another individual Chi Square works better for small sample sizes (< hundreds of samples) WARNING: Almost any really large table will have a significant chi square INFO 515 Lecture #8

Assumptions for Chi Square A random sample is the “expected” basis for comparison Each case can fall into only one cell No zero values are allowed for the observed frequency, fo And no expected frequencies, fe, less than one At least 80% of expected frequencies, fe, should be greater than or equal to five (≥5) INFO 515 Lecture #8

Expected Frequency The expected frequency for a cell is based on the fraction of things which would fall into it randomly, given the same general row and column count proportions as the actual data set fe = (row total) * (column total) / N So if 90 people live in New England, and 335 are in Age Group 1 from a total sample of 1500, then we would expect fe = 90*335/1500 = 20.1 people in that cell See slide 21 INFO 515 Lecture #8

Expected Frequency So the general formula for the expected frequency of a given cell is: fe = (actual row total)* (actual column total)/N Notice that this is NOT using the average expected frequency for every cell fe = N / [(# of rows)*(# of columns)] INFO 515 Lecture #8

Calculating Chi Square The Chi square value for each cell is the observed frequency minus the expected one, squared, divided by the expected frequency Chi square per cell = (fo-fe)2/fe Sum this for all cells in the crosstab For the cell on slide 28, the actual frequency was 25, so Chi square for that cell is = (25-20.1)2/20.1 = 1.195 Note: Chi square is always positive INFO 515 Lecture #8

Calculating Chi Square Page 36/37 of the Action Research handout has an example of chi square calculation, where fo is the observed (actual) frequency fe is the expected frequency E.g. fe for the first cell is 20*30/60 = 10.0 Chi square for each cell is (fo-fe)2/fe Sum chi square for all cells in the table No comments about fe fi fo fum! Is that clear?!?! INFO 515 Lecture #8

Interpreting Chi Square When the total Chi square is larger than the critical value, reject the null hypothesis See Action Research handout page 42/43 for critical Chi square (2) values Look up critical value using the ‘df’ value, which is based on the number of rows and columns in the crosstab: df = (#rows - 1)(#columns - 1) For the example on slide 21, df = (9-1)(4-1) = 8*3 = 24 INFO 515 Lecture #8

Interpreting Chi Square Or you can be lazy and use the old standby: if the significance is less than 0.050, reject the null hypothesis if the significance is less than 0.050, reject the null hypothesis if the significance is less than 0.050, reject the null hypothesis if the significance is less than 0.050, reject the null hypothesis INFO 515 Lecture #8

Notice we’re still using the Crosstab command! Chi Square Example Open data set “GSS91 political.sav” Use Analyze / Descriptive Statistics / Crosstabs... Set the Row(s) as “region”, and the Column(s) as “agegroup” Click on “Statistics…” and select the “Chi-square” test Notice we’re still using the Crosstab command! INFO 515 Lecture #8

Chi Square Example INFO 515 Lecture #8

Chi Square Example Note that we correctly predicted the ‘df’ value of 24 SPSS is ready to warn you if too many cells expected a count below five, or had expected counts below one The significance is below 0.050, indicating we reject the null hypothesis The total Chi square for all cells is 43.260 INFO 515 Lecture #8

Chi Square Example The critical Chi square value can be looked up on page 42/43 of Yonker For df = 24, and significance level 0.050, we get a critical Chi square of 36.415 Since the actual Chi square (43.260) is greater than the critical value (36.415), reject the null hypothesis Chi square often shows significance falsely for large sample sizes (hence the earlier warning) INFO 515 Lecture #8

Chi Square Example What are the other tests? They don’t apply here... The Likelihood Ratio test is specifically for log-linear models The Linear-by-Linear Association test is a function of Pearson’s ‘r’, so it only applies to interval or ratio scale variables Notice that SPSS doesn’t realize those tests don’t apply, and blindly presents results for them… INFO 515 Lecture #8

One-variable Chi square Test To check only one variable’s distribution, there is another way to run Chi square Null hypothesis is that the variable is evenly distributed across all of its categories Hence all expected frequencies are equal for each category, unless you specify otherwise Expected range can also be specified INFO 515 Lecture #8

Other Chi square Example Use Analyze / Nonparametric Tests / Chi-square… NOT using the Crosstab command here Add “region” to the Test Variable List Now df is the number of categories in the variable, minus one df = (# categories) - 1 Significance is interpreted the same INFO 515 Lecture #8

Other Chi square Example INFO 515 Lecture #8

Other Chi square Example So in this case, the “region” variable has nine categories, for a df of 9-1 = 8 Critical Chi square for df = 8 is 15.507, so the actual value of 290 shows these data are not evenly distributed across regions Significance below 0.050 still, in keeping with our fine long established tradition, rejects the null hypothesis INFO 515 Lecture #8

Whodunit? The chi-square value by itself doesn’t tell us which of the cells are major contributors to the statistical significance We compute the standardized residual to address that issue This hints at which cells contribute a lot to the total chi square INFO 515 Lecture #8

Residuals The Residual is the Observed value minus the Estimated value for some data point Residual = fo - fe If this variable is evenly distributed, the Residuals should have a normal distribution Plots of residuals are sometimes used to check data normalcy (i.e. how normal is this data’s distribution?) INFO 515 Lecture #8

Standardized Residual The Standardized Residual is the Residual divided by the standard deviation of the residuals When the absolute value of the Standardized Residual for a cell is greater than 2, you may conclude that it is a major contributor to the overall chi-square value Analogous to the original t test, looking for |t| > 2 INFO 515 Lecture #8

Standardized Residual Extreme values of Standardized Residual (e.g. minimum, maximum) can also help identify extreme data points The meaning of residual is the same for regression analysis, BTW, where residuals are an optional output INFO 515 Lecture #8

Standardized Residual Example In the crosstab region-agegroup example Click “Cells…” and select Standardized Residuals In this case, the worst cell is the combination W. Nor. Central region - Age Group 4, which produced a standardized residual of 2.1 INFO 515 Lecture #8

Standardized Residual Example INFO 515 Lecture #8

Crosstab Statistics for 2x2 Table 2x2 tables appear so often that many tests have been developed specifically for them Equality of proportions McNemar Chi-square Yates Correction Fisher Exact Test INFO 515 Lecture #8

Crosstab Statistics for 2x2 Table Equality of proportions tests prove whether the proportion of one variable is the same as for two different values of another variable e.g. Do homeowners vote as often as renters? McNemar Chi-square tests for frequencies in a 2x2 table where samples are dependent (such as pre-test and post-test results) INFO 515 Lecture #8

Crosstab Statistics for 2x2 Table Yates Correction for Continuity chi-square is refined for small observed frequencies fe = ( |fo-fe| - 0.5)/fe Corrections are too conservative; don’t use! Fisher Exact Test – assumes row/column frequencies remain fixed, and computes all possible tables; gives significance value like Chi square INFO 515 Lecture #8

Nominal Measures of Association Are used to test if each measure is zero (null hypothesis) using different scales Phi Cramer’s V Contingency Coefficient All three are zero iff Chi square is zero “iff” is mathspeak for ‘if and only if’ INFO 515 Lecture #8

Nominal Measures of Association The usual Significance criterion is used for all three If significance < 0.050, reject the null hypothesis, hence the association is significant Notice that direction is meaningless for nominal variables, so only the strength of an association can be determined INFO 515 Lecture #8

Phi For a 2x2 table, Phi and Cramer’s V are equal to Pearson’s r Phi (φ) can be > 1, making it an unusual measure of association Phi = sqrt[ (Chi square) / N] Phi = 0 means no association Phi near or over 1 means strong association INFO 515 Lecture #8

Cramer’s V Cramer’s V ≤ 1 V = sqrt[ Chi Square / (N*(k – 1) ] where k is the smaller of the number of columns or rows Is a better measure for tables larger than 2x2 instead of the Contingency Coefficient INFO 515 Lecture #8

Contingency Coefficient a.k.a. C or Pearson’s C or Pearson’s Contingency Coefficient Most widely used measure based on chi-square Requires only nominal data C has a value of 0 when there is no association INFO 515 Lecture #8

Contingency Coefficient The max possible value of C is the square root of (the number of columns minus 1, divided by the number of columns) Cmax = sqrt( (#column - 1) / #column) C = sqrt[ Chi Square / (Chi Square + N) ] INFO 515 Lecture #8