Presentation is loading. Please wait.

Presentation is loading. Please wait.

Inferential Statistics & Test of Significance

Similar presentations


Presentation on theme: "Inferential Statistics & Test of Significance"— Presentation transcript:

1 Inferential Statistics & Test of Significance

2 Confidence Interval (CI)
Y = mean Z = Z score related with a 95% CI σ = standard error

3 Building a CI Assume the following

4 CI Why do we use 1.96?

5 Is there a sample that is different from the mean?
Source; Knoke & Bohrnstead (1991:167) Is there a sample that is different from the mean?

6 Significance Testing When we explain some phenomenon we move beyond description to inferential statistics and hypothesis testing. Tests of significance allow us to test hypotheses, and when we find a relationship between variables, reject the null hypothesis.

7 Hypothesis testing Hypothesis testing means that we are testing our null hypothesis (Ho) against some competing or alternative hypothesis (H1) Normally we choose statements such as Ho : μy = 100 H1: μy ≠100 Or H1: μy > 100 H1: μy < 100

8 Significance Testing Even with high powered statistical measures, there will be results that pop up that are affected by chance. If we were to keep running our models a thousand times, or fewer, we would likely see some results that do not stem from systematic processes. Thus, we need to determine at what level of significance we are willing to frame our results. We can never be 100% confident. Conventional levels of significance where we reject the null hypothesis are usually .05 or The probability .10 is weakly significant.

9 Significance Testing When you erroneously reject the null hypothesis when it is true, you make a Type I error. This means you are accepting a “False Positive” result. Think of this as a fiancé test. The chances of rejecting or saying no to mister or miss “right”

10 Significance Testing A Type II error occurs when you accept the null hypothesis when it is not true. This is a “False Negative”, when you have say yes to Mr. or Miss “wrong” Type II errors in statistical testing result from too little data, omitted variable bias, and multicollinearity.

11 Other distributions The normal distribution assumes:
We know the standard error of the population, however, often we don’t know it. The t-distribution become the best alternative when we don’t know the standard error but we know the standard deviation. As the sample gets bigger the t-distribution approaches the normal distribution There are other distribution such as chi square and the that we will discuss latter.

12 T- Distribution & Normal Distribution
The form of the t-distribution depends on the sample size. As the sample gets Larger there is not difference between the normal and the t-distribution Source: Gujarati (1992:76)

13 The t formula For α =.05 and N=30 , t =2.045

14 95% CI using t-test 20± 2.093 (5/√20) = 22.34 upper 18.88 lower
Mean= 20 Sy = 5 N= 20 20± (5/√20) = 22.34 upper 18.88 lower

15 Why do we care about CI? We use CI interval for hypothesis testing
For instance, we want to know if there is a difference of home values between El Paso and Boston We want to know whether or not taking class at Kaplan makes a difference in our GRE scores We want to know if there is a difference between the treatment and control groups.

16 Mean Difference testing
Mean USA Las Cruces Boston El Paso Home Values

17

18 T-Tests of Independence
Used to test whether there is a significant difference between the means of two samples. We are testing for independence, meaning the two samples are related or not. This is a one-time test, not over time with multiple observations. Example: The values of homes between El Paso and Boston

19 T-Test of Independence
Useful in experiments where people are assigned to two groups, when there should be no differences, and then introduce Independent variables (treatment) to see if groups have real differences, which would be attributable to introduced X variable. This implies the samples are from different populations (with different μ). This is the Completely Randomized Two-Group Design.

20 T-Test of Independence
For example, we can take a random sample of high school students and divided into two groups. One gets tutoring for the SAT and the other does not. Ho: μ1≠ μ2 H1: μ1= μ2 After one group gets tutoring, but not the other, we compare the scores. We find that indeed the group exposed to tutoring outperformed the other group. We thus conclude that tutoring makes a difference.

21 Positive increments at a different rate
Treatment Control Post-test Pre-test

22 Two Sample Difference of Means T-Test
Pooled variance of the two groups Sp2 = = common standard deviation of two groups

23 Two Sample Difference of Means T-Test
The nominator of the equation captures difference in means, while the denominator captures the variation within and between each group. Important point: of interest is the difference between the sample means, not sample and population means. However, rejecting the null means that the two groups under analysis have different population means.

24 An example Test on GRE verbal test scores by gender:
Females: mean = 50.9, variance = , n=6 Males: mean=41.5, variance= , n=10

25 Now what do we do with this obtained value?

26 Steps of Testing and Significance
Statement of null hypothesis: if there is not one then how can you be wrong? Set Alpha Level of Risk: .10, .05, .01 Selection of appropriate test statistic: T-test, Computation of statistical value: get obtained value. Compare obtained value to critical value: done for you for most methods in most statistical packages.

27 Steps of Testing and Significance
Comparison of the obtained and critical values. If obtained value is more extreme than critical value, you may reject the null hypothesis. In other words, you have significant results. If point seven above is not true, obtained is lower than critical, then null is not rejected.

28 GRE Verbal Example Obtained Value: 2.605 Critical Value?
Degrees of Freedom: number of cases left after subtracting 1 for each sample. (14) Ho : μf =μm H1: μf ≠μm Is the null hypothesis (Ho) supported? Answer: No, women have higher verbal skills and this is statistically significant. This means that the mean scores of each gender as a population are different.

29 Paired T-Tests We use Paired T-Tests, test of dependence, to examine a single sample subjects/units under two conditions, such as pretest - posttest experiment. For example, we can examine whether a group of students improves if they retake the GRE exam. The T-test examines if there is any significant difference between the two studies. If so, then possibly something like studying more made a difference.

30 Paired T-Tests Unlike a test for independence, this test requires that the two groups/samples being evaluated are dependent upon each other. For example, we can use a paired t-test to examine two sets of scores across time as long as they come from the same students. This is appropriate for a pre-test –post-test research design

31 ΣD = sum differences between groups, plus it is squared.
n = number of paired groups

32 Comparing Test Scores Midterm Final 48 71.2 69 73.3 95 96 87 94.2 50
81.4 75 86.7 74 72.8 88 92 91.8 86 93.6 73 71.8 60 80.1

33 What can we conclude?


Download ppt "Inferential Statistics & Test of Significance"

Similar presentations


Ads by Google