Download presentation

Presentation is loading. Please wait.

Published byMarissa Sheehan Modified over 3 years ago

1
Descriptive Statistics-II Dr Mahmoud Alhussami

2
Shapes of Distribution A third important property of data – after location and dispersion - is its shape Distributions of quantitative variables can be described in terms of a number of features, many of which are related to the distributions physical appearance or shape when presented graphically. modality Symmetry and skewness Degree of skewness Kurtosis

3
Modality The modality of a distribution concerns how many peaks or high points there are. A distribution with a single peak, one value a high frequency is a unimodal distribution.

4
Modality A distribution with two or more peaks called multimodal distribution.

5
Symmetry and Skewness A distribution is symmetric if the distribution could be split down the middle to form two haves that are mirror images of one another. In asymmetric distributions, the peaks are off center, with a bull of scores clustering at one end, and a tail trailing off at the other end. Such distributions are often describes as skewed. When the longer tail trails off to the right this is a positively skewed distribution. E.g. annual income. When the longer tail trails off to the left this is called negatively skewed distribution. E.g. age at death.

6
Symmetry and Skewness Shape can be described by degree of asymmetry (i.e., skewness). mean > medianpositive or right-skewness mean = mediansymmetric or zero-skewness mean < mediannegative or left-skewness Positive skewness can arise when the mean is increased by some unusually high values. Negative skewness can arise when the mean is decreased by some unusually low values.

7
Left skewed: Right skewed: Symmetric: 7

8
12 January Shapes of the Distribution Three common shapes of frequency distributions: Three common shapes of frequency distributions: Symmetrical and bell shaped Positively skewed or skewed to the right Negatively skewed or skewed to the left ABC

9
12 January Shapes of the Distribution Three less common shapes of frequency distributions: Three less common shapes of frequency distributions: BimodalReverse J-shaped Uniform ABC

10
10 This guy took a VERY long time!

11
Degree of Skewness A skewness index can readily be calculated most statistical computer program in conjunction with frequency distributions The index has a value of 0 for perfectly symmetric distribution. A positive value if there is a positive skew, and negative value if there is a negative skew. A skewness index that is more than twice the value of its standard error can be interpreted as a departure from symmetry.

12
Measures of Skewness or Symmetry Pearson s skewness coefficient It is nonalgebraic and easily calculated. Also it is useful for quick estimates of symmetry. It is defined as: skewness = mean-median/SD Fisher s measure of skewness. It is based on deviations from the mean to the third power.

13
Pearson s skewness coefficient For a perfectly symmetrical distribution, the mean will equal the median, and the skewness coefficient will be zero. If the distribution is positively skewed the mean will be more than the median and the coefficient will be the positive. If the coefficient is negative, the distribution is negatively skewed and the mean less than the median. Skewness values will fall between -1 and +1 SD units. Values falling outside this range indicate a substantially skewed distribution. Hildebrand (1986) states that skewness values above 0.2 or below -0.2 indicate severe skewness.

14
Assumption of Normality Many of the statistical methods that we will apply require the assumption that a variable or variables are normally distributed. With multivariate statistics, the assumption is that the combination of variables follows a multivariate normal distribution. Since there is not a direct test for multivariate normality, we generally test each variable individually and assume that they are multivariate normal if they are individually normal, though this is not necessarily the case.

15
Evaluating normality There are both graphical and statistical methods for evaluating normality. Graphical methods include the histogram and normality plot. Statistical methods include diagnostic hypothesis tests for normality, and a rule of thumb that says a variable is reasonably close to normal if its skewness and kurtosis have values between – 1.0 and None of the methods is absolutely definitive.

16
Transformations When a variable is not normally distributed, we can create a transformed variable and test it for normality. If the transformed variable is normally distributed, we can substitute it in our analysis. Three common transformations are: the logarithmic transformation, the square root transformation, and the inverse transformation. All of these change the measuring scale on the horizontal axis of a histogram to produce a transformed variable that is mathematically equivalent to the original variable.

17
Types of Data Transformations for moderate skewness, use a square root transformation. For substantial skewness, use a log transformation. For sever skewness, use an inverse transformation.

18
Computing Explore descriptive statistics To compute the statistics needed for evaluating the normality of a variable, select the Explore… command from the Descriptive Statistics menu.

19
Adding the variable to be evaluated First, click on the variable to be included in the analysis to highlight it. Second, click on right arrow button to move the highlighted variable to the Dependent List.

20
Selecting statistics to be computed To select the statistics for the output, click on the Statistics… command button.

21
Including descriptive statistics First, click on the Descriptives checkbox to select it. Clear the other checkboxes. Second, click on the Continue button to complete the request for statistics.

22
Selecting charts for the output To select the diagnostic charts for the output, click on the Plots… command button.

23
Including diagnostic plots and statistics First, click on the None option button on the Boxplots panel since boxplots are not as helpful as other charts in assessing normality. Second, click on the Normality plots with tests checkbox to include normality plots and the hypothesis tests for normality. Third, click on the Histogram checkbox to include a histogram in the output. You may want to examine the stem-and-leaf plot as well, though I find it less useful. Finally, click on the Continue button to complete the request.

24
Completing the specifications for the analysis Click on the OK button to complete the specifications for the analysis and request SPSS to produce the output.

25
The histogram An initial impression of the normality of the distribution can be gained by examining the histogram. In this example, the histogram shows a substantial violation of normality caused by a extremely large value in the distribution.

26
The normality plot The problem with the normality of this variables distribution is reinforced by the normality plot. If the variable were normally distributed, the red dots would fit the green line very closely. In this case, the red points in the upper right of the chart indicate the severe skewing caused by the extremely large data values.

27
The test of normality Problem 1 asks about the results of the test of normality. Since the sample size is larger than 50, we use the Kolmogorov-Smirnov test. If the sample size were 50 or less, we would use the Shapiro-Wilk statistic instead. The null hypothesis for the test of normality states that the actual distribution of the variable is equal to the expected distribution, i.e., the variable is normally distributed. Since the probability associated with the test of normality is < is less than or equal to the level of significance (0.01), we reject the null hypothesis and conclude that total hours spent on the Internet is not normally distributed. (Note: we report the probability as <0.001 instead of.000 to be clear that the probability is not really zero.) The answer to problem 1 is false.

28
The assumption of normality script An SPSS script to produce all of the output that we have produced manually is available on the course web site. After downloading the script, run it to test the assumption of linearity. Select Run Script… from the Utilities menu.

29
Selecting the assumption of normality script First, navigate to the folder containing your scripts and highlight the NormalityAssumptionAndTransformations.SBS script. Second, click on the Run button to activate the script.

30
Specifications for normality script The default output is to do all of the transformations of the variable. To exclude some transformations from the calculations, clear the checkboxes. Third, click on the OK button to run the script. First, move variables from the list of variables in the data set to the Variables to Test list box.

31
The test of normality The script produces the same output that we computed manually, in this example, the tests of normality.

32
When transformations do not work When none of the transformations induces normality in a variable, including that variable in the analysis will reduce our effectiveness at identifying statistical relationships, i.e. we lose power. We do have the option of changing the way the information in the variable is represented, e.g. substitute several dichotomous variables for a single metric variable.

33
Fisher s Measure of Skewness The formula for Fisher s skewness statistic is based on deviations from the mean to the third power. The measure of skewness can be interpreted in terms of the normal curve A symmetrical curve will result in a value of 0. If the skewness value is positive, them the curve is skewed to the right, and vice versa for a distribution skewed to the left. A z-score is calculated by dividing the measure of skewness by the standard error for skewness. Values above or below are significant at the 0.05 level because 95% of the scores in a normal deviation fall between and from the mean. E.g. if Fisher s skewness= and st.err. =0.197 the z- score = 0.195/0.197 = 0.99

34
Kurtosis The distribution s kurtosis is concerns how pointed or flat its peak. Two types: Leptokurtic distribution (mean thin). Platykurtic distribution (means flat).

35
Kurtosis There is a statistical index of kurtosis that can be computed when computer programs are instructed to produce a frequency distribution For kurtosis index, a value of zero indicates a shape that is neither flat nor pointed. Positive values on the kurtosis statistics indicate greater peakedness, and negative values indicate greater flatness.

36
Fishers measure of Kurtosis Fisher s measure is based on deviation from the mean to the fourth power. A z-score is calculated by dividing the measure of kurtosis by the standard error for kurtosis.

37
Table of descriptive statistics To answer problem 2, we look at the values for skewness and kurtosis in the Descriptives table. The skewness and kurtosis for the variable both exceed the rule of thumb criteria of 1.0. The variable is not normally distributed. The answer to problem 2 if false.

38
Other problems on assumption of normality A problem may ask about the assumption of normality for a nominal level variable. The answer will be An inappropriate application of a statistic since there is no expectation that a nominal variable be normal. A problem may ask about the assumption of normality for an ordinal level variable. If the variable or transformed variable is normal, the correct answer to the question is True with caution since we may be required to defend treating an ordinal variable as metric. Questions will specify a level of significance to use and the statistical evidence upon which you should base your answer.

39
Normal Distribution Also called belt shaped curve, normal curve, or Gaussian distribution. A normal distribution is one that is unimodal, symmetric, and not too peaked or flat. Given its name by the French mathematician Quetelet who, in the early 19 th century noted that many human attributes, e.g. height, weight, intelligence appeared to be distributed normally.

40
12 January Normal Distribution The normal curve is unimodal and symmetric about its mean (). The normal curve is unimodal and symmetric about its mean (). In this distribution the mean, median and mode are all identical. In this distribution the mean, median and mode are all identical. The standard deviation () specifies the amount of dispersion around the mean. The standard deviation () specifies the amount of dispersion around the mean. The two parameters and completely define a normal curve. The two parameters and completely define a normal curve.

41
Also called a Probability density function. The probability is interpreted as "area under the curve." The random variable takes on an infinite # of values within a given interval The probability that X = any particular value is 0. Consequently, we talk about intervals. The probability is = to the area under the curve. The area under the whole curve = 1. 41

42
42

43
Normal Distribution X is the random variable. μ is the mean value. σ is the standard deviation (std) value. e = constant. π = constant.

44
Importance of Normal Distribution to Statistics Although most distributions are not exactly normal, most variables tend to have approximately normal distribution. Many inferential statistics assume that the populations are distributed normally. The normal curve is a probability distribution and is used to answer questions about the likelihood of getting various particular outcomes when sampling from a population.

45
Probabilities are obtained by getting the area under the curve inside of a particular interval. The area under the curve = the proportion of times under identical (repeated) conditions that a particular range of values will occur. Characteristics of the Normal distribution: It is symmetric about the mean μ. Mean = median = mode. [ bell-shaped curve] f(X) decreases as X gets farther and farther away from the mean. It approaches horizontal axis asymptotically: - < X < +. This means that there is always some probability (area) for extreme values. 45

46
12 January Why Do We Like The Normal Distribution So Much? There is nothing special about standard normal scores There is nothing special about standard normal scores These can be computed for observations from any sample/population of continuous data values These can be computed for observations from any sample/population of continuous data values The score measures how far an observation is from its mean in standard units of statistical distance The score measures how far an observation is from its mean in standard units of statistical distance But, if distribution is not normal, we may not be able to use Z-score approach. But, if distribution is not normal, we may not be able to use Z-score approach.

47
12 January Probability Distributions Any characteristic that can be measured or categorized is called a variable. Any characteristic that can be measured or categorized is called a variable. If the variable can assume a number of different values such that any particular outcome is determined by chance it is called a random variable. If the variable can assume a number of different values such that any particular outcome is determined by chance it is called a random variable. Every random variable has a corresponding probability distribution. Every random variable has a corresponding probability distribution. The probability distribution applies the theory of probability to describe the behavior of the random variable. The probability distribution applies the theory of probability to describe the behavior of the random variable.

48
12 January Discrete Probability Distributions Binomial distribution – the random variable can only assume 1 of 2 possible outcomes. There are a fixed number of trials and the results of the trials are independent. Binomial distribution – the random variable can only assume 1 of 2 possible outcomes. There are a fixed number of trials and the results of the trials are independent. i.e. flipping a coin and counting the number of heads in 10 trials. i.e. flipping a coin and counting the number of heads in 10 trials. Poisson Distribution – random variable can assume a value between 0 and infinity. Poisson Distribution – random variable can assume a value between 0 and infinity. Counts usually follow a Poisson distribution (i.e. number of ambulances needed in a city in a given night) Counts usually follow a Poisson distribution (i.e. number of ambulances needed in a city in a given night)

49
12 January Discrete Random Variable A discrete random variable X has a finite number of possible values. The probability distribution of X lists the values and their probabilities. A discrete random variable X has a finite number of possible values. The probability distribution of X lists the values and their probabilities. 1. Every probability p i is a number between 0 and The sum of the probabilities must be 1. Find the probabilities of any event by adding the probabilities of the particular values that make up the event. Find the probabilities of any event by adding the probabilities of the particular values that make up the event. Value of X x1x1x1x1 x2x2x2x2 x3x3x3x3… xkxkxkxk Probability p1p1p1p1 p2p2p2p2 p3p3p3p3… pkpkpkpk

50
12 January Example The instructor in a large class gives 15% each of A s and D s, 30% each of B s and C s and 10% F s. The student s grade on a 4-point scale is a random variable X (A=4). The instructor in a large class gives 15% each of A s and D s, 30% each of B s and C s and 10% F s. The student s grade on a 4-point scale is a random variable X (A=4). What is the probability that a student selected at random will have a B or better? What is the probability that a student selected at random will have a B or better? ANSWER: P (grade of 3 or 4)=P(X=3) + P(X=4) ANSWER: P (grade of 3 or 4)=P(X=3) + P(X=4) = = 0.45 = = 0.45 GradeF=0D=1C=2B=3A=4 Probability

51
12 January Continuous Probability Distributions When it follows a Binomial or a Poisson distribution the variable is restricted to taking on integer values only. When it follows a Binomial or a Poisson distribution the variable is restricted to taking on integer values only. Between two values of a continuous random variable we can always find a third. Between two values of a continuous random variable we can always find a third. A histogram is used to represent a discrete probability distribution and a smooth curve called the probability density is used to represent a continuous probability distribution. A histogram is used to represent a discrete probability distribution and a smooth curve called the probability density is used to represent a continuous probability distribution.

52
12 January Normal Distribution Q Is every variable normally distributed? A Absolutely not Q Then why do we spend so much time studying the normal distribution? A Some variables are normally distributed; a bigger reason is the Central Limit Theorem !!!!!!!!!!!!!!!!!!!!!!!!!!!?????????? ?

53
Central Limit Theorem describes the characteristics of the "population of the means" which has been created from the means of an infinite number of random population samples of size (N), all of them drawn from a given "parent population". It predicts that regardless of the distribution of the parent population: The mean of the population of means is always equal to the mean of the parent population from which the population samples were drawn. The standard deviation of the population of means is always equal to the standard deviation of the parent population divided by the square root of the sample size (N). The distribution of means will increasingly approximate a normal distribution as the size N of samples increases.

54
Central Limit Theorem A consequence of Central Limit Theorem is that if we average measurements of a particular quantity, the distribution of our average tends toward a normal one. In addition, if a measured variable is actually a combination of several other uncorrelated variables, all of them "contaminated" with a random error of any distribution, our measurements tend to be contaminated with a random error that is normally distributed as the number of these variables increases. Thus, the Central Limit Theorem explains the ubiquity of the famous bell-shaped "Normal distribution" (or "Gaussian distribution") in the measurements domain.

55
Note that the normal distribution is defined by two parameters, μ and σ. You can draw a normal distribution for any μ and σ combination. There is one normal distribution, Z, that is special. It has a μ = 0 and a σ = 1. This is the Z distribution, also called the standard normal distribution. It is one of trillions of normal distributions we could have selected. 55

56
12 January Standard Normal Variable It is customary to call a standard normal random variable Z. It is customary to call a standard normal random variable Z. The outcomes of the random variable Z are denoted by z. The outcomes of the random variable Z are denoted by z. The table in the coming slide give the area under the curve (probabilities) between the mean and z. The table in the coming slide give the area under the curve (probabilities) between the mean and z. The probabilities in the table refer to the likelihood that a randomly selected value Z is equal to or less than a given value of z and greater than 0 (the mean of the standard normal). The probabilities in the table refer to the likelihood that a randomly selected value Z is equal to or less than a given value of z and greater than 0 (the mean of the standard normal).

57
57 Source: Levine et al, Business Statistics, Pearson.

58
12 January The Rule for the Normal Distribution 68% of the observations fall within one standard deviation of the mean 68% of the observations fall within one standard deviation of the mean 95% of the observations fall within two standard deviations of the mean 95% of the observations fall within two standard deviations of the mean 99.7% of the observations fall within three standard deviations of the mean 99.7% of the observations fall within three standard deviations of the mean When applied to real data, these estimates are considered approximate! When applied to real data, these estimates are considered approximate!

59
Remember these probabilities (percentages): Practice: Find these values yourself using the Z table. Two Sample Z Test 59 # standard deviations from the mean Approx. area under the normal curve ±1.68 ± ± ±2.955 ± ±3.997

60
12 January Standard Normal Curve

61
12 January Standard Normal Distribution 50% of probability in here–probability=0.5

62
12 January Standard Normal Distribution 2.5% of probability in here 95% of probability in here Standard Normal Distribution with 95% area marked

63
12 January Calculating Probabilities Probability calculations are always concerned with finding the probability that the variable assumes any value in an interval between two specific points a and b. Probability calculations are always concerned with finding the probability that the variable assumes any value in an interval between two specific points a and b. The probability that a continuous variable assumes the a value between a and b is the area under the graph of the density between a and b. The probability that a continuous variable assumes the a value between a and b is the area under the graph of the density between a and b.

64
Normal Distribution64 If the weight of males is N.D. with μ=150 and σ=10, what is the probability that a randomly selected male will weigh between 140 lbs and 155 lbs? [Important Note: Always remember that the probability that X is equal to any one particular value is zero, P(X=value) =0, since the normal distribution is continuous.]

65
Solution: Z = (140 – 150)/ 10 = s.d. from mean Area under the curve =.3413 (from Z table) Z = (155 – 150) / 10 =+.50 s.d. from mean Area under the curve =.1915 (from Z table) Answer: =

66
Example For example: What s the probability of getting a math SAT score of 575 or less, =500 and =50? i.e., A score of 575 is 1.5 standard deviations above the mean Yikes! But to look up Z= 1.5 in standard normal chart (or enter into SAS) no problem! =.9332

67
If IQ is ND with a mean of 100 and a S.D. of 10, what percentage of the population will have (a) IQs ranging from 90 to 110? (b) IQs ranging from 80 to 120? Solution: Z = (90 – 100)/10 = Z = ( )/ 10 = Area between 0 and 1.00 in the Z-table is.3413; Area between 0 and is also.3413 (Z-distribution is symmetric). Answer to part (a) is =

68
(b) IQs ranging from 80 to 120? Solution: Z = (80 – 100)/10 = Z = ( )/ 10 = Area between =0 and 2.00 in the Z-table is.4772; Area between 0 and is also.4772 (Z-distribution is symmetric). Answer is =

69
Suppose that the average salary of college graduates is N.D. with μ=$40,000 and σ=$10,000. (a) What proportion of college graduates will earn $24,800 or less? (b) What proportion of college graduates will earn $53,500 or more? (c) What proportion of college graduates will earn between $45,000 and $57,000? (d) Calculate the 80th percentile. (e) Calculate the 27th percentile. 69

70
(a) What proportion of college graduates will earn $24,800 or less? Solution: Convert the $24,800 to a Z-score: Z = ($24,800 - $40,000)/$10,000 = Always DRAW a picture of the distribution to help you solve these problems. 70

71
First Find the area between 0 and in the Z-table. From the Z table, that area is Then, the area from to - is = Answer: 6.43% of college graduates will earn less than $24, $24,800 $40, X Z

72
(b) What proportion of college graduates will earn $53,500 or more? Solution: Convert the $53,500 to a Z-score. Z = ($53,500 - $40,000)/$10,000 = Find the area between 0 and in the Z- table:.4115 is the table value. When you DRAW A PICTURE (above) you see that you need the area in the tail: Answer: Thus, 8.85% of college graduates will earn $53,500 or more. 72 $40,000 $53,500 Z

73
(c) What proportion of college graduates will earn between $45,000 and $57,000? Z = $45,000 – $40,000 / $10,000 =.50 Z = $57,000 – $40,000 / $10,000 = 1.70 From the table, we can get the area under the curve between the mean (0) and.5; we can get the area between 0 and 1.7. From the picture we see that neither one is what we need. What do we do here? Subtract the small piece from the big piece to get exactly what we need. Answer: = $40k Z $45k$57k

74
Parts (d) and (e) of this example ask you to compute percentiles. Every Z-score is associated with a percentile. A Z-score of 0 is the 50 th percentile. This means that if you take any test that is normally distributed (e.g., the SAT exam), and your Z-score on the test is 0, this means you scored at the 50 th percentile. In fact, your score is the mean, median, and mode. 74

75
(d) Calculate the 80 th percentile. Solution: First, what Z-score is associated with the 80 th percentile? A Z-score of approximately +.84 will give you about.3000 of the area under the curve. Also, the area under the curve between - and 0 is Therefore, a Z-score of +.84 is associated with the 80th percentile. Now to find the salary (X) at the 80 th percentile: Just solve for X: +.84 = (X$40,000)/$10,000 X = $40,000 + $8,400 = $48, $40,000 Z ANSWER

76
(e) Calculate the 27 th percentile. Solution: First, what Z-score is associated with the 27 th percentile? A Z-score of approximately -.61will give you about.2300 of the area under the curve, with.2700 in the tail. (The area under the curve between 0 and -.61 is.2291 which we are rounding to.2300). Also, the area under the curve between 0 and is Therefore, a Z-score of -.61 is associated with the 27 th percentile. Now to find the salary (X) at the 27 th percentile: Just solve for X: =(X$40,000)/$10,000 X = $40,000 - $6,100 = $33, $40,000 Z ANSWER.2700

77
12 January T-Distribution Similar to the standard normal in that it is unimodal, bell- shaped and symmetric. Similar to the standard normal in that it is unimodal, bell- shaped and symmetric. The tail on the distribution are thicker than the standard normal The tail on the distribution are thicker than the standard normal The distribution is indexed by degrees of freedom (df). The distribution is indexed by degrees of freedom (df). The degrees of freedom measure the amount of information available in the data set that can be used for estimating the population variance (df=n-1). The degrees of freedom measure the amount of information available in the data set that can be used for estimating the population variance (df=n-1). Area under the curve still equals 1. Area under the curve still equals 1. Probabilities for the t-distribution with infinite df equals those of the standard normal. Probabilities for the t-distribution with infinite df equals those of the standard normal.

78
12 January Graphical Methods Frequency Distribution Frequency Distribution Histogram Histogram Frequency Polygon Frequency Polygon Cumulative Frequency Graph Cumulative Frequency Graph Pie Chart.

79
Presenting Data Table Condenses data into a form that can make them easier to understand; Shows many details in summary fashion; BUT Since table shows only numbers, it may not be readily understood without comparing it to other values.

80
Principles of Table Construction Don t try to do too much in a table Us white space effectively to make table layout pleasing to the eye. Make sure tables & test refer to each other. Use some aspect of the table to order & group rows & columns.

81
Principles of Table Construction If appropriate, frame table with summary statistics in rows & columns to provide a standard of comparison. Round numbers in table to one or two decimal places to make them easily understood. When creating tables for publication in a manuscript, double-space them unless contraindicated by journal.

82
12 January Frequency Distributions A useful way to present data when you have a large data set is the formation of a frequency table or frequency distribution. A useful way to present data when you have a large data set is the formation of a frequency table or frequency distribution. Frequency – the number of observations that fall within a certain range of the data. Frequency – the number of observations that fall within a certain range of the data.

83
83 Frequency Table Age Number of Deaths < , , , , Total Total34,524

84
84 Frequency Table Data Intervals Frequency Frequency Cumulative Frequency Relative Frequency (%) Cumulative Relative Frequency (%) Total Total

85
12 January Cumulative Relative Frequency Cumulative Relative Frequency – the percentage of persons having a measurement less than or equal to the upper boundary of the class interval. Cumulative Relative Frequency – the percentage of persons having a measurement less than or equal to the upper boundary of the class interval. i.e. cumulative relative frequency for the 3 rd interval of our data example: i.e. cumulative relative frequency for the 3 rd interval of our data example: = 59.6% = 59.6% - We say that 59.6% of the children have weights below 39.5 pounds.

86
12 January Number of Intervals There is no clear-cut rule on the number of intervals or classes that should be used. There is no clear-cut rule on the number of intervals or classes that should be used. Too many intervals – the data may not be summarized enough for a clear visualization of how they are distributed. Too many intervals – the data may not be summarized enough for a clear visualization of how they are distributed. Too few intervals – the data may be over- summarized and some of the details of the distribution may be lost. Too few intervals – the data may be over- summarized and some of the details of the distribution may be lost.

87
Presenting Data Chart -Visual representation of a frequency distribution that helps to gain insight about what the data mean. - Built with lines, area & text: bar charts Ex: bar chart, pie chart

88
Bar Chart Simplest form of chart Used to display nominal or ordinal data

89
Horizontal Bar Chart

90
Cluster Bar Chart

91
Pie Chart Alternative to bar chart Circle partitioned into percentage distributions of qualitative variables with total area of 100%

92
Histogram Appropriate for interval, ratio and sometimes ordinal data Similar to bar charts but bars are placed side by side Often used to represent both frequencies and percentages Most histograms have from 5 to 20 bars

93
Histogram

94
12 January Pictures of Data: Histograms Blood pressure data on a sample of 113 men Blood pressure data on a sample of 113 men Histogram of the Systolic Blood Pressure for 113 men. Each bar spans a width of 5 mmHg on the horizontal axis. The height of each bar represents the number of individuals with SBP in that range.

95
12 January Frequency Polygon First place a dot at the midpoint of the upper base of each rectangular bar. The points are connected with straight lines. At the ends, the points are connected to the midpoints of the previous and succeeding intervals (these intervals have zero frequency).

96
Hallmarks of a Good Chart Simple & easy to read Placed correctly within text Use color only when it has a purpose, not solely for decoration Make sure others can understand chart; try it out on somebody first Remember: A poor chart is worse than no chart at all.

97
12 January Cumulative Frequency Plot Place a point with a horizontal axis marked at the upper class boundary and a vertical axis marked at the corresponding cumulative frequency. Each point represents the cumulative relative frequency and the points are connected with straight lines. The left end is connected to the lower boundary of the first interval that has data.

98
The Uses of Frequency Distributions Becoming familiar with dataset. Cleaning the data. Outliers-values that lie outside the normal range of values for other cases. Inspecting the data for missing values. Testing assumptions for statistical tests. Assumption is a condition that is presumed to be true and when ignored or violated can lead to misleading or invalid results. When DV is not normally distributed researchers have to choose between three options: Select a statistical test that does not assume a normal distribution. Ignore the violation of the assumption. Transform the variable to better approximate a distribution that is normal. Please consult the various data transformation.

99
The Uses of Frequency Distributions Obtaining information about sample characteristics. Directing answering research questions.

100
Outliers Are values that are extreme relative to the bulk of scores in the distribution. They appear to be inconsistent with the rest of the data. Advantages: They may indicate characteristics of the population that would not be known in the normal course of analysis. Disadvantages: They do not represent the population Run counter to the objectives of the analysis Can distort statistical tests.

101
Sources of Outliers An error in the recording of the data. A failure of data collection, such as not following sample criteria (e.g. inadvertently admitting a disoriented patient into a study), a subject not following instructions on a questionnaire, or equipment failure. An actual extreme value from an unusual subjects.

102
Methods to Identify Outliers Traditional way of labeling outliers, any value more than 3SD from the mean. Values that are more than 3 IQRs from the upper or lower edge of the box plot are extreme outliers.

103
Handling Outliers Analyze the data two ways: With the outliers in the distribution With outliers removed. If the results are similar, as they are likely to be if the sample size is large, then the outliers may be ignored. If the results are not similar, then a statistical analysis that is resistant to outliers can be used (e.g. median and IQR). If you want to use a mean with outliers, then the trimmed mean is an option. If calculated with a certain percentage of the extreme values removed from both ends of the distribution (e.g. n=100, then 5% trimmed mean is the mean of the middle 90% of the observation).

104
Handling Outliers Another alternative is a Winsorized mean. The highest and lowest extremes are replaced by the next-to-highest value and by the next-to-lowest value. For Univariate outliers, Tabachnick and Fidell (2001) suggest changing the scores on the variables for the outlying cases so they are deviant. E.g. if the two largest scores in the distribution are 125 and 122 and the next largest score 87. recode 122 as 88 and 125 as 89.

105
Missing Data Any systematic event external to the respondent (such as data entry errors or data collection problems) or action on the part of the respondent (such as refusal to answer) that leads to missing data. It means that analyses are based on fewer study participants than were in the full study sample. This, in turn, means less statistical power, which can undermine statistical conclusion validity-the degree to which the statistical results are accurate. Missing data can also affect internal validity-the degree to which inferences about the causal effect of the dependent variable on the dependent variable are warranted, and also affect the external validity-generalizability.

106
Strategies to avoid Missing Data Persistent follow-up Flexibility in scheduling appointments Paying incentives. Using well-proven methods to track people who have moved. Performing a thorough review of completed data forms prior to excusing participants.

107
Techniques for Handling Missing Data Deletion techniques. Involve excluding subjects with missing data from statistical calculation. Imputation techniques. Involve calculating an estimate of each missing value and replacing, or imputing, each value by its respective estimate. Note: techniques for handling missing data often vary in the degree to which they affect the amount of dispersion around true scores, and the degree of bias in the final results. Therefore, the selection of a data handling technique should be carefully considered.

108
Deletion Techniques Deletion methods involve removal of cases or variables with missing data. Listwise deletion. Also called complete case analysis. It is simply the analysis of those cases for which there are no missing data. It eliminates an entire case when any of its items/variables has a missing data point, whether or not that data point is part of the analysis. It is the default of the SPSS. Pairwise deletion. Called the available case analysis (unwise deletion). Involves omitting cases from the analysis on a variable-by-variable basis. It eliminates a case only when that case has missing data for variables or items under analysis. Note: deletion techniques are widely criticized because they assume that the data are MCAR (which is very difficult to ascertain), pose a risk for bias, and lead to reduction of sample size and power.

109
Imputation Techniques Imputation is the process of estimating missing data based on valid values of other variables or cases in the sample. The goal of imputation is to use known relationship that can be identified in the valid values of the sample to help estimate the missing data

110
Types of Imputation Techniques Using prior knowledge. Inserting mean values. Using regression.

111
Prior Knowledge Involves replacing a missing value with a value based on an educational guess. It is a reasonable method if the researcher has a good working knowledge of the research domain, the sample is large, and the number of missing values is small.

112
Mean Replacement Also called median replacement for skewed distribution. Involves calculating mean values from a available data on that variable and using them to replace missing values before analysis. It is a conservative procedure because the distribution mean as a whole does not change and the researcher does not have to guess at missing values.

113
Using Regression Involves using other variables in the dataset as independent variables to develop a regression equation for the variable with missing data serving as the dependent variable. Cases with complete data are used to generate the regression equation. The equation is then used to predict missing values for incomplete cases. More regressions are computed, using the predicted values from the previous regression to develop the next equation, until the predicted values from one step to the next are comparable. Prediction from the last regression are the ones used to replace missing values.

114

Similar presentations

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google