Chi-Square and Nonparametric Techniques

Slides:



Advertisements
Similar presentations
What is Chi-Square? Used to examine differences in the distributions of nominal data A mathematical comparison between expected frequencies and observed.
Advertisements

Chapter 18: The Chi-Square Statistic
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Hypothesis Testing Steps in Hypothesis Testing:
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Chi square.  Non-parametric test that’s useful when your sample violates the assumptions about normality required by other tests ◦ All other tests we’ve.
Parametric/Nonparametric Tests. Chi-Square Test It is a technique through the use of which it is possible for all researchers to:  test the goodness.
Chi Square Tests Chapter 17.
Hypothesis Testing IV Chi Square.
 What is chi-square  CHIDIST  Non-parameteric statistics 2.
S519: Evaluation of Information Systems
Independent Sample T-test Formula
PSY 340 Statistics for the Social Sciences Chi-Squared Test of Independence Statistics for the Social Sciences Psychology 340 Spring 2010.
Statistics Are Fun! Analysis of Variance
PY 427 Statistics 1Fall 2006 Kin Ching Kong, Ph.D Lecture 12 Chicago School of Professional Psychology.
CHI-SQUARE GOODNESS OF FIT TEST u A nonparametric statistic u Nonparametric: u does not test a hypothesis about a population value (parameter) u requires.
Ch 15 - Chi-square Nonparametric Methods: Chi-Square Applications
Intro to Statistics for the Behavioral Sciences PSYC 1900 Lecture 17: Chi-Square.
CHI-SQUARE statistic and tests
PSY 307 – Statistics for the Behavioral Sciences Chapter 19 – Chi-Square Test for Qualitative Data Chapter 21 – Deciding Which Test to Use.
Richard M. Jacobs, OSA, Ph.D.
Inferential Statistics
Chapter Ten Introduction to Hypothesis Testing. Copyright © Houghton Mifflin Company. All rights reserved.Chapter New Statistical Notation The.
AM Recitation 2/10/11.
Hypothesis Testing:.
Statistics for the Behavioral Sciences (5 th ed.) Gravetter & Wallnau Chapter 17 The Chi-Square Statistic: Tests for Goodness of Fit and Independence University.
Overview of Statistical Hypothesis Testing: The z-Test
Week 10 Chapter 10 - Hypothesis Testing III : The Analysis of Variance
Which Test Do I Use? Statistics for Two Group Experiments The Chi Square Test The t Test Analyzing Multiple Groups and Factorial Experiments Analysis of.
Chapter 11 HYPOTHESIS TESTING USING THE ONE-WAY ANALYSIS OF VARIANCE.
Chapter 9: Non-parametric Tests n Parametric vs Non-parametric n Chi-Square –1 way –2 way.
1 Chi-Square Heibatollah Baghi, and Mastee Badii.
Chi-squared Tests. We want to test the “goodness of fit” of a particular theoretical distribution to an observed distribution. The procedure is: 1. Set.
Chapter 16 The Chi-Square Statistic
Nonparametric Tests: Chi Square   Lesson 16. Parametric vs. Nonparametric Tests n Parametric hypothesis test about population parameter (  or  2.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
1 Chapter 8 Introduction to Hypothesis Testing. 2 Name of the game… Hypothesis testing Statistical method that uses sample data to evaluate a hypothesis.
Fitting probability models to frequency data. Review - proportions Data: discrete nominal variable with two states (“success” and “failure”) You can do.
CHI SQUARE TESTS.
© aSup-2007 CHI SQUARE   1 The CHI SQUARE Statistic Tests for Goodness of Fit and Independence.
© Copyright McGraw-Hill 2004
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Chapter 13 Understanding research results: statistical inference.
HYPOTHESIS TESTING FOR DIFFERENCES BETWEEN MEANS AND BETWEEN PROPORTIONS.
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
Chi Square 11.1 Chi Square. All the tests we’ve learned so far assume that our data is normally distributed z-test t-test We test hypotheses about parameters.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Cross Tabulation with Chi Square
Chapter 9: Non-parametric Tests
Lecture Nine - Twelve Tests of Significance.
Research Methodology Lecture No :25 (Hypothesis Testing – Difference in Groups)
INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE
Inference and Tests of Hypotheses
Math 4030 – 10b Inferences Concerning Variances: Hypothesis Testing
Hypothesis Testing Review
Community &family medicine
John Loucks St. Edward’s University . SLIDES . BY.
Qualitative data – tests of association
Inferential Statistics
Chapter 9 Hypothesis Testing.
Part IV Significantly Different Using Inferential Statistics
Reasoning in Psychology Using Statistics
Quantitative Methods in HPELS HPELS 6210
What are their purposes? What kinds?
Reasoning in Psychology Using Statistics
Inference for Two Way Tables
Chapter Nine: Using Statistics to Answer Questions
Chapter 10 Introduction to the Analysis of Variance
Chapter 18: The Chi-Square Statistic
Presentation transcript:

Chi-Square and Nonparametric Techniques Q560: Experimental Methods in Cognitive Science Lecture 14

The Tolman/Hull Debate (1946) Can animals create “cognitive maps” or are they purely stimulus-response learners? This was a huge debate in the 40s between Clark Hull (S-R) and Edward Tolman (cognition). In a classic experiment, Tolman trained rats to run down an alley into a circular arena. From there, they were trained to go straight across to another alley which turned right to a goal box After training, he put them in a new maze that also contained alleys at other angles, one heading directly towards the goal box. The rats had a choice to take the alley they had learned, or a novel one that pointed in the direction of the goal

If Hull was right, the rat learned an S-R sequence during learning, and would continue this response If Tolman was right, the rat learned a cognitive map “goal on the right” and would enter the alley on the right Goal B A C D Start

If Hull was right, the rat learned an S-R sequence during learning, and would continue this response If Tolman was right, the rat learned a cognitive map “goal on the right” and would enter the alley on the right Goal Alley Chosen B A C A B C D Observed Expected 4 8 5 15 D Who was right? Start

Chi-Square Test Our first problem is that each rat has only provided us with an alley preference…this is a nomial scale (frequency) not an interval or ratio scale, so we cannot compute means or variances So, we use a Chi-Square statistic: comparing observed frequencies to those expected by chance Why divide by expected freq? Relative discrepancy from expected: Party example

Expected Frequencies A B C D 25% A B C D 8 If the null hypothesis is true, alley choices should be random…no preference. So with 4 options: p(option) = .25 A B C D 25% For each option, , where p = proportion if null is true, and n is sample size (n=32 rats) A B C D 8 This is our chance model

Chi-Square Test As with our other statistical tests, we compare to a chance distribution for Chi-square. Like the F-distribution, we actually have a family of distributions depending on df (Table on p. 697) df = C -1 (where C = # of columns)...why?

The shape can change quite a bit depending on df

For Tolman’s rats: df = C - 1 = 4-1 = 3 For df = 3, X2-crit = 7.81 Note: df are based on the number of “options” rather than sample size (as with our previous tests). Hence, as the number of options increases, the chance of a difference due purely to chance also increases. So, the critical value for Chi-square actually increases with larger df rather than decreasing (as with t and F) Alley Chosen A B C D Observed Expected 4 8 5 15

Alley Chosen X2-crit = 7.81 X2-obt = 9.25 A B C D 4 8 5 15 Observed Expected 4 8 5 15 X2-crit = 7.81 X2-obt = 9.25

Steps in Hypothesis Testing State hypotheses (No preference/preference) Determine critical values (chance model) Calculate Chi-sqaure statistic Decision and conclusions

An Example Coke Pepsi RC Cola 13 9 8 We’ll settle the age-old debate of whether people can actually detect their favorite cola based solely on taste. For 30 coke-lovers, I blindfold them, and have them sample 3 colas…is there a true difference, or are these preference differences explainable by chance? Coke Pepsi RC Cola 13 9 8

Step 1: Hypotheses Null: There are no preferences: The population is divided evenly among the brands Alternate: There are preferences: The population is not divided evenly among the brands Note: This is non-parametric, so we don’t need to talk about parameters (our “code symbols”)

Step 2: Chance Model Coke Pepsi RC Cola Observed 13 9 8 Expected 10 df = C -1 = 3 -1 = 2, set  = .05 For df = 2, X2-crit = 5.99 Next, we determine the frequencies we would expect if the null hypothesis were true ....is this 1- or 2-tailed? Coke Pepsi RC Cola Observed 13 9 8 Expected 10

Step 3: Calculate Chi-Square Coke Pepsi RC Cola Observed 13 9 8 Expected 10

Step 4: Decision and Conclusion Conclude that the preferences are evenly divided among the colas when the logos are removed.

An SPSS example w/ RPS

The Chi-Square Test of Independence This is the nominal equivalent of a factorial ANOVA: we have two variables, and we want to test whether there is a relationship between them. We present the data in a matrix. If there is no relationship, then the observed frequencies should not differ as a function of the second variable

Littering 17 28 49 Not Littering 73 62 41 Cialdini et al. (1990) examined whether a person’s tendency to litter depended on the amount of litter already in the area. People were handed a flyer as they entered an amusement park. The entrance area had already been prepared with either no litter, a small amount of litter, or a lot of litter lying on the ground. The frequency data are: Amount of Litter Small Amount Large Amount None Littering 17 28 49 Not Littering 73 62 41 Is there a relationship between tendency to litter and litter already in the area?

Littering 17 28 49 Not Littering 73 62 41 94 176 90 90 90 n = 270 Amount of Litter Small Amount Large Amount None Littering 17 28 49 Not Littering 73 62 41 94 176 90 90 90 n = 270 Where: df = (R-1)(C-1) = (3-1)(2-1) = 2

Litt (Obs): Expected: 17 31.33 28 49 No Litt (Obs): 73 58.67 62 41 94 Amount of Litter Small Amount Large Amount None Litt (Obs): Expected: 17 31.33 28 49 No Litt (Obs): 73 58.67 62 41 94 176 90 90 90 n = 270 Where:

There is a relationship between tendency to litter and litter already in the area

Likelihood Ratios Note that the likelihood ratio is chi-square distributed with k-1 degrees of freedom: In model testing, the likelihood ratio is used to determine the likelihood of Model A generating the data over Model B. In the chi-square case, we are determining the likelihood that the observed data are generated by a chance model.