# Statistics BLITZ!!! By Gina M. Salvati *~DISCLAIMER~*

## Presentation on theme: "Statistics BLITZ!!! By Gina M. Salvati *~DISCLAIMER~*"— Presentation transcript:

Statistics BLITZ!!! By Gina M. Salvati *~DISCLAIMER~*
Some examples are lifted from Fundamentals of Statistics Third Edition. All rights reserved, yada, yada, yada. I don’t own those examples, and they are noted throughout the presentation. The only people benefitting from this presentation (hopefully) are peer tutors and the students. ENJOY!

First, and foremost… Don’t Panic Unlike Algebra, Statistics is not an exact math – there will always be some kind of variation among answers. So if your answer doesn’t exactly match what is in the text or (especially) on the answer key on an exam, don’t panic. As long as you are in the same ballpark as the given answer(s), you should be just fine.  Even with that being said, however, always, always, always check your work. It is always important to double-check yourself, no matter which course you take. A good test-taking tip when it comes to multiple choice is to go with the answer closest to what you come up with. Yes, that was a Hitchhiker’s Guide to the Galaxy reference. The alien, too. XD

The Chamber of Doom (I Mean, Table of Contents) Types of Curves
The Empirical Rule Measures of Central Tendency vs. Measures of Dispersion What to Use and When to Use it {Pt.1}: normalcdf vs. invNorm How to Calculate on TI-83/84 The Sampling Distribution of 𝑥 vs. The Sampling Distribution of 𝑝 Z-scores The Five-Number Summary, Boxplots, and Outliers What to Use and When to Use it {Pt. 2}: Confidence Intervals To Construct a Boxplot What to Use and When to Use it (Pt. 3}: The Insanity that is Hypothesis Testing The Interquartile Range To Determine Outliers Feat. Type I and Type II Errors Example What to Use and When to Use it {The Finale}: Matched-Pair Data vs. 𝜒 2 −GOF Testing Linear Regression and What All Goes With It r and r2 A Quick Look at Probability

The Types of Curves Left-skew Symmetrical AKA Bell Curve Right-skew
Mean = Median = Mode Mean < Median < Mode Mean > Median > Mode In the image above, the relationships of mean, median, and mode are illustrated.

Measures of Central Tendency vs. Measures of Dispersion
The Measure of Central Tendency basically means the average or typical data value. For this class, it mostly refers to where our distribution is centered at. There are 3 Measures of Central Tendency: Mean The measure we mostly refer to when we say “the average.” Two types: population (𝜇) and sample ( 𝑥 ) Median The value that lies in the center of our data when it is all put in ascending order. The most resistant measure of central tendency – i.e., it’s not affected by extreme values (meaning it doesn’t get yanked around by outliers and crazy numbers as much as, say, the mean) Mode The number(s) in our data that occur(s) most frequently.

Measures of Central Tendency vs. Measures of Dispersion
The Measure of Dispersion refers to what we use to describe the spread of our distribution. Like the Measures of Central Tendency, there are three of these: Range Highest number – lowest number The Interquartile Range (IQR) falls under this Variance (standard deviation)2 Two types: population ( 𝜎 2 ) and sample ( 𝑠 2 ) Standard Deviation The most common form of dispersion and the one that we use in this course Two types: population (𝜎) and sample (𝑠)

How to Calculate on TI-83/84
Example Time! Flight Time The following data represent the mean flight time (in minutes) of a random sample of seven flights from Las Vegas, Nevada, to Newark, New Jersey, on Continental Airlines: 282, 270, 260, 266, 257, 260, 267 Calculate the mean, median, and mode flight time. Enter data in L1 (Stat, 1:Edit…) Stat, Calc, 1: 1-Var Stats ENTER The list that appears gives you all the information you will be looking for. 𝑥 provides the mean, 𝑠𝑥 provides the sample standard deviation (s.d.), and 𝜎𝑥 provides the population s.d. Scroll down until you see Med. This number is your median. The only thing it doesn’t calculate is the mode (the most reoccurring number), but that’s pretty easy to find on its own. ;) From Fundamentals of Statistics Third Edition text, pg. 125.

Z-scores The z-score basically represents the distance that a data value is from the mean in terms of the number of standard deviations; i.e., it’s a converted number. Sounds a lot more complicated than it looks. All you need to know is one simple equation: 𝑧 𝑠𝑐𝑜𝑟𝑒 = 𝑠𝑐𝑜𝑟𝑒−𝑚𝑒𝑎𝑛 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑑𝑒𝑣𝑖𝑎𝑡𝑖𝑜𝑛 Let’s see an example…

Z-scores Example: Men vs. Women The average 20- to 29-year-old man is 69.6 inches tall, with a standard deviation of 3.0 inches, while the average 20- to 29-year-old woman is 64.1 inches tall, with a standard deviation of 3.8 inches. Who is relatively taller, a 67-inch man or a 62-inch woman? Break the information into two (2) equations and then compare: Men: 𝑧= 67 − = − ≈−0.867 Women: 𝑧= 62 − = − ≈−0.553 The relatively taller person is… The 67-inch man. From Fundamentals of Statistics Third Edition text, pg. 161.

The Five-Number Summary, Boxplots, and Outliers
The Five-Number Summary basically divvies up your data into quartiles. It also can be used to construct a boxplot. The Five-Number Summary consists of the following: Smallest number in the data set (Minimum) First quartile ( 𝑄 1 ) Median ( 𝑄 2 ) Third quartile ( 𝑄 3 ) Largest number in the data set (Maximum) To get it on your calculator, enter your data in L1 under Stat, 1: Edit…, and then proceed as if you were looking for the mean, median, and mode (Stat, Calc, 1: 1-Var Stats). Scroll down to the bottom of the list, and you will find the Five-Number Summary. ;)

To Construct a Boxplot After entering your data in L1, hit 2nd, Y=. This will bring you to the Stat Plots screen. Hit ENTER or 1. This will bring you to another screen, where you can actually set up the boxplot. Turn the plots ON. Select the Boxplot with Outliers type. This is the first image in the second row in the Types category. Zoom, 9: ZoomStat Ta-da! :D

The Interquartile Range
The Interquartile Range (IQR) is the range the middle 50% of observations in the data set. Again, another simple little equation: 𝐼𝑄𝑅= 𝑄 3 − 𝑄 1

To Determine OutLiers Outliers are basically extreme values – the numbers that are waaay out there on either side of the data set when all the data are put in ascending order. To determine if there are any outliers in a data set, we use the following two equations: Lower Fence – any numbers below this number are considered outliers 𝐿𝐹 = 𝑄 1 −1.5(𝐼𝑄𝑅) Upper Fence – any numbers above this number are considered outliers 𝑈𝐹= 𝑄 (𝐼𝑄𝑅)

The Five-Number Summary, Boxplots, and Outliers
EXAMPLE TIME! April Showers The following data represent the number of inches of rain in Chicago, Illinois, during the month of April for 20 randomly selected years. [a] Determine the five-number summary [b] Compute the Interquartile Range (IQR) [c] Determine the Upper and Lower Fences. Are there any outliers? 0.97 2.47 3.94 4.11 5.79 1.14 2.78 3.97 4.77 6.14 1.85 3.41 4.00 5.22 6.28 2.34 4.02 5.50 7.69 Fundamentals of Statistics Third Edition, pg. 162

The Five-Number Summary, Boxplots, and Outliers – Answers
[a] Determine the quartiles Record all data in L1 Stat, Calc: 1:1-Var Stats Min: .97 𝑄 1 : 2.405 𝑄 2 /med: 3.985 𝑄 3 : 5.36 Max: 7.69 [b] Compute the IQR Q3 – Q1 = IQR 5.36 – = 2.955 [c] Determine the Upper and Lower Fences. Are there any outliers? LF = Q1 – 1.5(IQR) LF = – 1.5(2.955) LF = UF = Q (IQR) UF = (2.955) UF = Because there are no numbers in the data set that are less than the lower fence or greater than the upper fence, there are no outliers.

Linear Regression and What All Goes with It
Linear Regression is basically a method of modeling the relationship between the independent/explanatory variable (x) and the dependent variable (y). The simplest way to look at it is to think in terms of equation of lines. The least-squares regression line equation is pretty much slope-intercept form – no joke. To solve on TI-83/84: Enter data in Stat, 1: Edit… X values in L1, Y values in L2 Stat, Calc, 4: LinReg As shown in the image on the right (), this screen will appear. If you do not see r and r2, you will need to set it up: 2nd, 0 (Catalog), scroll down to DiagnosticOn (or hit the 𝑥 −1 key to save time), ENTER Hit ENTER again; a message saying DONE will appear. ;) Look at this equation setup here. It should look familiar. (Hint: substitute m for a)  What is r and r2? Well, let’s take a look…

R and r2 The linear correlation coefficient (r) measures how closely related the data is to the linear regression line The coefficient of determination (r2) holds explaining power: “the linear regression explains __ much of the data.” Note: It is important to remember that correlation doesn’t necessarily imply causation – just because there is a correlation between two variables, it doesn’t always mean that one is causing the other.

A Quick Look at Probability
Probability measures the likelihood of something happening (and, in some cases, not happening). Like everything else in Statistics, there are a few rules: The probability of any event (E) must be greater than or equal to 0 and less than or equal to 1, AKA: 0≤𝑃 𝐸 ≤1 The sum of all probabilities must equal to one If an event is impossible, its probability is 0. Conversely, if an event is certain, its probability is 1. An unusual event has a less than 5% (.05) chance of occurring. This is an important rule to remember, as it holds true in several other sections in this course.

The Empirical Rule The Empirical Rule is used to give us an approximation of the number(s) that lies within 1, 2, or 3 standard deviations away from the mean. 68% of the distribution lies within 1 deviation from the mean 95% of the distribution lies within 2 deviations of the mean 99.7% of the distribution lies within 3 deviations of the mean

The Empirical Rule: A Slightly better Example
= 110 95 – 5 = 90 68% 95% 99.7% 𝜇 =100 90 – 5 = 85 = 115 100 – 5 = 95 = 105 𝜎=5

What to Use and When to Use It {Pt. 1}: normalcdf vs. invNorm
Two simple calculator functions that can very easily get mixed up. These are the key things to remember about normalcdf and invNorm: normalcdf: When we’re given z-scores or data points* and asked to find the probability/proportion/area under the curve. invNorm: When we’re given a percent/probability/area and asked to find the z-score/data point. A trick is to look for key words. One example is the word percentile. If a question asks you to find a percentile, you will use the invNorm function on your calculator.  * Z-scores are what we use when dealing with the standard normal curve (𝜇=0, 𝑠=1); when working with a normal distribution (such is in the example in slides 25-26), we refer to the numbers as data points.

Normalcdf on the Calculator
When using the normalcdf function on your calculator, the order of input is the lower bound, followed by the upper bound, then the mean, and finally the standard deviation. Most calculators will have you enter the information manually; others are kind enough to help you out a little: -E99 and E99 are very useful in circumstances when you are not given an upper or lower bound. In some instance, -100 and 100 can be used, but sometimes a much smaller or larger may be called for. To put this in the calculator, press 2nd and then the comma key to get E. Simply type the negative symbol before if doing –E99. 

Normalcdf Example If I were to ask what the area under the curve was above 1.50, my graph would look something like this: If were to ask what the area under the curve was below 1.50, my graph would look like this: And if I wanted to know the area between, say, and 1.5, my graph would look like this: NOTE: In this example, we are working with the standard normal curve. This means that μ = 0 and σ = 1.

invNorm on The Calculator
When using the invNorm function on your calculator, the order of input is the area to the left of the z-score we’re seeking, followed by the mean, and finally the standard deviation. Most calculators will have you enter the information manually; others are kind enough to help you out a little:

invNorm Example 1.0 invNorm problems can get a little confusing at times. The key, as with virtually everything in Statistics, is in the wording. Let’s say I am in charge of a chocolate chip cookie factory. The factory churns out an average of 3500 cookies a day, with a standard deviation of 45. Inventory is just around the corner, and we’re expected to make above the 60th percentile. Over how many cookies are we supposed to make to meet expectations? The wording here isn’t exactly the greatest, but we are given the information necessary to perform the task. We know that we a have an average (mean, μ) of 3500 cookies and a standard deviation (σ) of 45. We also know that we need to make above the 60th percentile. This means we need to make over the first 60% in order to meet the quota. How do we solve for this?

invNorm Example 1.0 (cont.)
The easiest way to solve is to start out by drawing a picture: The 60th percentile is illustrated as the first 60% of our distribution. Using some more of our newly discovered mad calculator skills, we can find the number of cookies (our data point) by following these steps: 2nd, vars, 3: invNorm Enter, in order, the area to the left side of the data point we are seeking (here, .60), the mean, and the standard deviation ENTER The number you get will be the data point we’re looking for. The answer here comes out to be Because we are talking about cookies and not random numbers, it is best to round. A good practice is to round up, to take into account for potential extra. ANSWER: In order to meet expectations, the factory needs to make over 3511/3512 cookies.

The Sampling Distribution of 𝑥 vs. the Sampling Distribution of 𝑝
A simple process gone wrong, or at least this is what it sometimes feels like with sampling distributions. Actually, sampling distributions are really nifty little paths that help us when dealing with samples and proportions. There are two that we deal with: The Sampling Distribution of 𝑥 and the Sampling Distribution of 𝑝 . The difference? Let’s find out…

The Sampling Distribution of the Sample Mean
The sampling distribution of the sample mean (AKA: x-bar, 𝑥 ) is used when we are given information about a sample and asked to find the probability of selecting individuals possessing a certain characteristics we are looking for. The sampling distribution goes a little something like this: The distribution is approximately normal 𝜇 𝑥 = 𝜇 𝜎 𝑥 = 𝜎 𝑛 This cute little equation is called the standard error of the mean, or just standard error. Good little vocab hint to remember. ;) Let’s see an example…

Example Problem #1: Sampling Distribution of 𝑥
A simple random sample size n = 49 is obtained from a population with µ = 80 and σ = 14. Describe the sampling distribution of 𝑥 . What is 𝑃 𝑥 >83 ? What is 𝑃 𝑥 ≤75.8 ? What is 𝑃(59.8 < 𝑥 <65.9)? From Fundamentals of Statistics, Third Edition, pg. 389

Example Problem #1 Solution
The sampling distribution of x-bar is approximately normal. This is pretty much always the case, at least for this course. If the sample is not normal, we cannot proceed with the sampling distribution. To determine if a sample is normal, the general rule of thumb is to look at your sample size (n). If the sample size is 30 or greater (n ≥ 30), you can go ahead; if the sample size is smaller than 30, the information must specify that the population is normal. The rest of the sampling distribution is as follows: 𝜇 𝑥 = 𝜇=80 𝜎 𝑥 = 𝜎 𝑛 = = =2

Example Problem #1 Solution (cont.)
The sampling distribution is approximately normal, with 𝜇 𝑥 = 80 and 𝜎 𝑥 = 2. To solve, it’s best to start out by drawing a “pretty picture” before plugging numbers in the calculator: For each part, select 2nd, vars, 2: normalcdf. This operation will give you the area under the curve.  First curve: 2nd, vars, 2; (83, E99, 80,2) ≈ .0668 Second curve: 2nd, vars, 2; (-E99, 75.8, 80, 2) ≈ .0179 Third curve: 2nd, vars, 2: ( Please note: images aren’t exactly to scale. ^^;;

The Sampling Distribution of the Population Proportion
The sampling distribution of the population proportion (AKA p-hat or 𝑝 ) is used when we are given information about a sample and a proportion/percent and asked to find the probability of selecting individuals possessing a certain characteristics we are looking for. This sampling distribution looks like this: The distribution is approximately normal 𝜇 𝑝 =𝑝 𝜎 𝑝 = 𝑝(1−𝑝) 𝑛 𝑝 = 𝑥 𝑛 This equation is for a point estimate. We use this to adjust our numbers in order to fit the distribution. If you’re dealing with decimals, a whole number isn’t going to fall in your distribution. ;)

Example Problem: Sampling Distribution of 𝑝
Dun Dun Duuunnn!!! Smith owns a shipyard. He knows that 5% of all welding done that afternoon will wind up being defective. Out of the 7000 welds in the yard, he examines What’s the probability that between 10 and 20 welding jobs will be defective?

Smith and His Shipyard This kind of a question is a classic nuisance. First of all, how do we determine what numbers represent our sample (n) and the number(s) out of it (x)? Look at the information. The problem states that out of 7000 welds in the yard, Smith examines 300 welds. 300 is our sample size; the 7000 total welds is population in the shipyard. Out of the 300 welds he’s examined, he wants to know the probability between 10 and 20 welding jobs that could be defective. 10 and 20 represent our x values. Knowing this, we can now move forward with our information gathering: p = 5% = .05 n = 300 𝑥 1 = 10 𝑥 2 = 20

Smith and His Shipyard (Cont.)
We can now go ahead with our sampling distribution: 𝜇 𝑝 = p = .05 𝜎 𝑝 = (enter eq on 2010) = .0126 𝑥 1 =10/300 = .033 𝑥 2 = 20/300 = .067 Finally, using our mad calculator skills, we can determine the area between our two boundaries (aka – z-scores): 2nd, vars, 2 (LB, UB, 𝜇 𝑝 , 𝜎 𝑝 ) (.033, .067, .05, .0126) TA DA!!! The answer is approximately .8227

What to Use and When to Use It {Pt. 2}: Confidence Intervals
A Quick Run-Down Stat, Tests,… 7: Z-Int – when given σ (population standard deviation) 8: T-Int – when given s (sample standard deviation) or no standard deviation when given a data set A: 1-PropZInt – when given x and n values (a sample, n, and a number out the sample, x) and a percent or proportion B: 2-PropZInt – when given two x and n values (two samples and numbers out of those samples: x1 and n1, and x2 and n2)

What to Use and When to Use it {Pt
What to Use and When to Use it {Pt. 3}: The Insanity that is Hypothesis Testing A Quick Run-Down Stat, Tests,… 1: Z-Test – when given σ (population standard deviation) 2: T-Test – when given s (sample standard deviation) or no standard deviation when given a data set 5: 1-PropZTest – when given x and n values (a sample, n, and a number out of the sample, x) and a percent or proportion 6: 2-PropZTest – when given two x and n values (two samples and numbers out of those samples: x1 and n1, and x2 and n2)

Feat. Type I and Type II Errors
The best way to describe these two is to think about hypothesis testing as a court case. The old motto of “innocent until proven guilty” provides us with our null and alternative hypotheses. The following is a mini-cartoon illustrating Type I and Type II errors.

What to Use and When to Use It {The Finale}: Matched-Pair Data vs
What to Use and When to Use It {The Finale}: Matched-Pair Data vs. Chi-Square GOF Testing Now, this is where things are confusing. Because these two types of testing appear so alike when entering data into the calculator, it’s easy to get confused. How can we tell the difference between these two kinds of problems??? WHAT TO LOOK FOR, HOW TO SOLVE The problems from these two sections appear to be very similar, but they are very different. It’s important to remember when to use either method of solving. Let’s take a look at a couple of examples…

Matched-Pair Data On one final exam study guide, there is a problem that talks about a football coach claiming that players can increase their strength by taking a certain supplement. He decides to test this theory by randomly selecting 9 athletes and gives them a strength test on the bench press. 30 days later, after regular training and taking this supplement, he tests them again. First of all, we need to identify what kind of problem this is. I just identified it as a matched-pair data problem. How can you tell if it’s this and not chi-square goodness- of-fit? Look at the data table: We want to test if the coach’s claim holds – that the supplement is effective in increasing the athletes’ strength. Athlete 1 2 3 4 5 6 7 8 9 Before 215 240 188 212 275 260 225 200 185 After 245 210 282 230 195 190

Matched-Pair Data The “Before” values are the first set of data we gathered. These will go into List 1 (Stat, Edit, 1). The values in the second data set we gathered (“After”) will go into List 2. Okay, now we need to see how the difference between these two data sets levels out in our test against our level of significance, α = To do this, we first need to get the differences, and we do that by highlighting the List 3 heading and entering the equation L1 – L2 and pressing the ENTER key. This will give you a list of numbers – the difference between List 1 and List 2. Exit the screen by pressing 2nd, mode. Then, go to Stat, Tests, 2 (Z-Test), and use the DATA option for your input. Change the List selection from L1 to L3, and proceed accordingly with your hypothesis testing. In short: L1 – “before” values L2 – “after” values L3 heading – L1 – L2 2nd, mode (to get out) Stat, Tests, 2 (Z-Test) Data Look at L3 **Note**: When writing out your null and alternate hypotheses, remember the H0 will always be μ = 0, to indicate that there is no difference in values. To determine H1, look at the alternative claim.

Remember: Problems from Matched-Pair Data and Chi-Square may look somewhat similar, but they are very different. The confusion lies in 1.) identifying the problem and 2.) remembering which equation to put in the L3 heading. * *If you have a TI-84 calculator, there is a trick to avoid this issue. The trick will be covered in another slide.

Chi-Square Again, problems from this section and from matched-pair data may look somewhat similar, but they are very different. Here is an example: The information provided in one problem from a former exam review gives us the number of wins for track hurdlers and asks us to test the claim that “the probabilities of winning are the same in different positions.” Again, we have to identify whether this is actually chi-square and not matched-pair data. Look at the chart, which I’ve added to here as well: Unlike the previous question, which gave us “before” and “after” values, this gives us a table of observed outcomes. This is where we can tell what the difference is: matched-pair data are generally given as “before” and “after,” whereas chi-square gives us an expectation and what we actually observed happening. Now that we know this much, what is our expectation for this equation? Like any other hypothesis test, we have to identify a null hypothesis and an alternative hypothesis. In this case, our null is what we expected. For a problem like this, we would expect that, if a runner each lane had equal chances of winning, each lane would get a 1/6 chance. Starting Position 1 2 3 4 5 6 Number of Wins 45 50 36 44 32 33

Chi-Square (Cont) How does this work on a calculator? Similar to matched-pair data problems, we start out by entering data (Stat, Edit, 1). In L1, we put our observed data – what actually happened. This would be the number of wins represented in the table above. In L2, we take the total number of wins (here, 240) and multiply it by the chance we expected (1/6). And in the L3 heading, we use the equation (L1 – L2)2/L2. In short: L1 – observed outcomes (what really happened); number of wins in graph L2 – expected outcomes (what we thought was going to happen); total wins * chance of winning (250 * 1/6) L3 heading – (L1 – L2)2/L2 2nd, mode (to get out) 2nd, Stat, < , Math, 5 (sum) Find the sum of L3 This is your test statistic Proceed accordingly with hypothesis testing

Chi-Square (cont.) Another way to find the chi-square test statistic is as follows: L1 – observed outcomes (what really happened) L2 – expected outcomes (what we thought was going to happen) 2nd, Mode (to get out) Stat, Tests, D (alpha, 𝑥 −1 key): 𝜒 2 GOF-Test You will see the following on your screen: Enter your degrees of freedom (df) and hit Calculate This will give you not only the test statistic, but it will give you the p-value as well.  Website for detailed walkthrough on how to enter on calculator: https://epsstore.ti.com/OA_HTML/csksxvm.jsp?nSetId=93701 (PLEASE NOTE: this trick only works on TI-84 calculators; TI-83 calculators do not have this function!)

You Ain’t Never Had a Friend Like...
The Ultimate Wrap-Up for What to Use and When to Use it! :D Given Hypothesis Testing Confidence Intervals σ (population standard deviation) 1 7 s (sample standard deviation) or no standard deviation when given a data set 2 8 When given x and n values (a sample, n, and a number out of the sample, x) 5 A When given x and n values for two samples (x1 and n1, x2 and n2) 6 B

This concludes the Statistics Blitz PowerPoint Presentation 
THE End! This concludes the Statistics Blitz PowerPoint Presentation  Created by Gina M. Salvati, CRLA certified tutor and former peer tutor for North Florida Community College and Florida Gateway College Version updated:

Download ppt "Statistics BLITZ!!! By Gina M. Salvati *~DISCLAIMER~*"

Similar presentations