Presentation is loading. Please wait.

Presentation is loading. Please wait.

Chapter 10 Discriminant Analysis

Similar presentations


Presentation on theme: "Chapter 10 Discriminant Analysis"— Presentation transcript:

1 Chapter 10 Discriminant Analysis
Introduction Discriminant analysis linear equation Fisher’s discriminant analysis Bayes discriminant analysis Stepwise discriminant analysis Discriminant analysis in SPSS(version 20) Summary Glossary Exercises

2 Statistics in practice
When a company, for example WAHAHA, is going to produce a new kind of product , at the beginning, the managers usually want to predict whether this product will be a success or not. There are many factors affecting the future of this new product, such as consumption level and consumption habits of the potential consumers, infrastructure of the country, the wrapping, design and quality of the product etc. They need to do a market survey to find the data of these factors. Relying on the dataset, how to predict whether this new product will be a success or not? One way is through the discriminant analysis.

3 10.1 Introduction Discriminant Analysis (DA) Scope Of DA DA Steps
DA Classification

4 Discriminant Analysis (DA)
A set of tools and methods used to distinguish between two or more groups of populations and to determine how to allocate new observations into groups. Discriminant Analysis plays the same role as the multiple linear regressions by predicting an outcome. The basic purpose of discriminant analysis is to estimate the relationship between a single categorical dependent variable and a set of quantitative independent variables.

5 Scope Of DA Discriminant analysis has widespread applications in situations where the primary objective is identifying the group to which a categorical object belongs. Include predicting the success or failure of a new product, determining what category of credit risk a person falls into, classifying students as to vocational interests, deciding whether a student should be admitted to graduate school or predicting whether a company will be a success or not.

6 DA Steps Discriminant analysis is usually splitted into a 2-step process. The first step is to test the significance of a set of discriminant functions. The second step is the classification.

7 DA Classification Two-group discriminant analysis and multi-group discriminant analysis by the number of discriminant groups. linear discriminant analysis and nonlinear discriminant analysis according to the used mathematical models. step discriminant analysis and sequential discriminant analysis according to the different methods used for variables Fisher discriminant analysis and Bayes discriminant analysis according to different discriminant criterions.

8 10.2 Discriminant analysis linear equation
Discriminant Analysis Form Main Assumptions

9 Discriminant Analysis Form
The linear combination for a discriminant analysis, also known as the discriminant analysis function, is derived from an equation that takes the following form: where is the discriminate function; is the coefficient or weight of that variable; is a variable and is a constant.

10 Main Assumptions The sample is a random one and each predictor variable is normally distributed. The allocations for the dependent categories in the initial classification are correctly classified. There are two or more categorical groups, and each case belongs to exactly one group, so that the groups are mutually exclusive and collectively exhaustive. Each group must be well defined, clearly differentiated from any other group. Putting a median split on an attitude scale is not a natural way to form groups. Partitioning quantitative variables is only justifiable if there are easily identifiable gaps at the points of division.

11 Main Assumptions The groups should be defined before we collect the data. The attributes used to separate the groups should discriminate quite clearly between the groups. Hence there is no group overlap or the group overlap is minimal. Group sizes of the dependent should not be extremely different.

12 10.3 Fisher’s discriminant analysis
The idea of Fisher’s discriminant analysis One-way ANOVA Rule

13 The idea of Fisher’s discriminant analysis
The Fisher’s discriminant analysis was raised by R. Fisher in 1936. The idea of Fisher’s discriminant analysis is to base the discriminant rule on a projection, such that a good separation of groups of -dimensional data was achieved.

14 One-way ANOVA Rule. Let denote a linear combination of observations, where is a vector of a space , then the total sum of squares of , , is equal to (10.1) With the centering matrix and . Assume we have samples , , where is a positive integer denoting the sample size. Then find the linear combination which maximizes the ratio of the between-group-sum of squares to the within-group-sum of squares.

15 One-way ANOVA Rule. The within-group-sum of squares is given by (10.2) Where denotes the -th submatrix of corresponding to observations the -th group, and denotes the centering matrix. The within-group-sum of squares measures the sum of variations within each group.

16 One-way ANOVA Rule. The between-group-sum of squares is (10.3) Where and denote the means of and , and and denote the means of and , respectively. The between-group-sum of squares measures the variation of the means across groups. The total sum of squares (10.1) is the sum of the within-group-sum of squares and the between-group-sum of squares, that is,

17 One-way ANOVA Rule. Fisher’s discriminant analysis selects the projection vector that maximizes the ration (10.4) The solution can be easily found. Since the vector which maximizes formula (10.4) is the eigenvector of hat corresponds to the largest eigenvalue, a discriminant rule is easy to obtain. Classifying into group where is closest to . For more properties and details of Fisher’s discriminant analysis, we refer to [1,2,3].

18 10.4 Bayes discriminant analysis
The idea of Bayes discriminant analysis Introduce The Principle

19 The idea of Bayes discriminant analysis
This kind of information can usually be utilized to form a prior probability distribution. According to the principle of Bayes’ rule, the prior distribution can possibly be further refined after obtaining the observations, and refined to form the posterior distribution. The discriminant involving the posterior distribution is often referred to as Bayes discriminant.

20 Introduce The Principle
Let . Thus the ECM (expected cost of misclassification) caused by the above discriminant is: The Bayes discriminant analysis is to find a classification of such that ECM is minimum. For more Bayes discriminant analysis properties and detailed mathematical proofs refer to [1,2,3].

21 Introduce The Principle
Assume that each set has a density function where . We already know that the prior probability distributions are respectively. Then Suppose that a classification of a space is given: . Hence whenever and . The discriminant rule is usually the following form: if belongs to , We use denoting the cost where the sample coming from but was wrongly sent to , and the probability of this misjudgment is

22 Introduce The Principle
Let . Thus the ECM (expected cost of misclassification) caused by the above discriminant is: The Bayes discriminant analysis is to find a classification of such that ECM is minimum. For more Bayes discriminant analysis properties and detailed mathematical proofs refer to [1,2,3].

23 10.5 Stepwise discriminant analysis
The idea of Stepwise discriminant analysis The procedures of undertaking stepwise discriminant analysis in SPSS

24 The idea of Stepwise discriminant analysis
Stepwise discriminant analysis is an attempt to find the best set of predictors. It is often used in an exploratory situation to identify those variables from among a larger number that might be used later in a more rigorous theoretically driven study. In stepwise discriminant analysis, the most correlated independent is usually entered first by the stepwise program, then the second, until an additional dependent adds no significant amount to the canonical R squared.

25 The procedures of undertaking stepwise discriminant analysis in SPSS
The most economical method is the Wilks Lambda method. Hit “Analyze”, then choose “Classify”, then select “Discriminant”. Select grouping variable and transfer to “Grouping Variable” box. Then hit “Define Range” button and enter the lowest and highest codes for your grouping variable define range. Hit “Continue” then choose predictors and enter into “Independents box”. Then click on “Use Stepwise Methods”. Choose “Statistics”, then “Means, Univariate Anovas, Box’s M, Unstandardized and Within Groups Correlation”. Hit “Classify”. Select “Compute From Group Sizes, Summary Table, Leave One Out Classifi cation, Within Groups”, and all “Plots”. Click“Continue”, then “Save” and select “Predicted Group Membership” and “Discriminant Scores”.

26 10.6 Discriminant analysis in SPSS(version 20)
Example 1 Example 2

27 Example 1 To undertake discriminant analysis in SPSS, in the “Analyze” menu, click the “Classify”, then click “Discriminant”. In this case we are looking at a dataset that describes mobile phones with a drawback A. Research tells us that the degree of symptom drawback A in mobile phones is often in the eye of the beholder—some people may think that drawback A affects the mobile phone, and other people may not think it as a problem. For example women may think that drawback affects the mobile phone, while men may think that the drawback A does not affect the mobile phone.

28 Example 1 We are going to look at whether the symptoms reported on four different measures ( , , ,and ) can tell us whether those symptom ratings were provided by a woman (“people” = 1) or by a man (“people” = 2). In the dialog box, put in “people” as the “grouping variable” (in other words, the variable that you think defines different groups). It will appear in the box with two question marks after it—you have to tell SPSS what the codes are for the two groups that you want to compare. Click “Define Range” and type in “1” and “2” as the different values of the “people” variable that you want to compare. Then click “Continue” to go on.

29 Example 1 Then put the four “question” variables ( , ,
and ) in as the predictors (“independents”). The sub-dialog box “statistics” lets you see descriptive statistics on each predictor variable for the different groups. Let’s check “Means” to see some basic descriptive statistics.

30 Example 1

31 Example 1

32 Example 1 Click “Continue” to go back to the main dialog box. Then, in the main dialog box, click the “Classify” button to see the next sub-dialog. Here’s the sub-dialog that you get when you hit the “Classify” button:

33 Example 1

34 Example 1 On this screen, check the box for the “summary table.” This will give you the classification table (sensitivity, specificity, etc.) on your printout. Hit “Continue” to go back to the main dialog box, and then hit “OK” to see the results of your analysis.

35 Example 1 This table just tells you if there’s any missing data.
The following is the result. Discriminant [DataSet1] C:\Users\lenovo\Documents\peopleexample.sav This table just tells you if there’s any missing data.

36 Example 1 This table shows the means that we asked for—it gives means on each variable for people in each sub-group, and also the overall means on each variable.

37 Example 1 Analysis 1 Summary of Canonical Discriminant Functions
This table tells you something about the “latent variable” that you have constructed (the discriminant function), which helps you to differentiate between the groups.

38 Example 1 Here is the multivariate test—Wilks’ lambda, just like in MANOVA.

39 Example 1 These “discriminant function coefficients” work just like the beta-weights in regression. Based on these, you can write out the equation for the discriminant function: Using this equation, given someone’s scores on , , and , you can calculate their score on the discriminant function. To figure out what that D function score means, look at the following group centroids.

40 Example 1 This table tells you the correlation between each item and the discriminant function, but we won’t use it for anything here.

41 Example 1 In practical terms, we usually figure out which group a person is in by calculating a cut score halfway between the two centroids: If an individual person’s score on the function (calculated by plugging in their scores on , , , and to the equation we wrote out above) is above –1.034, then they were probably men. If their function score is below –1.034, then they were probably women

42 Example 1 Classification Statistics

43 Example 1 Here is the classification table that we got by selecting that option in the SPSS dialog box. It gives information about actual group membership and predicted group membership.

44 Example 2 A sample of 12 riding-lawnmower owners and 12 nonowners is sampled from a city and the income in thousands of dollars and lot size in thousands of square feet are recorded. A riding-mower manufacturer wants to see if these two variables adequately separate owners from nonowners, and if so then direct their marketing on the basis of the separation of owners from nonowners.

45 Example 2 The data table is the following. Lawnmower owners and nonowners dataset

46 Example 2 If a family has a lawnmower, then ; if a family does not have a lawnmower, then . In the dialog box, put in “family” as the “grouping variable” (the variable that you think defines different groups). It will appear in the box with two question marks after it. Click “Define Range” and type in “1” and “2” as the different values of the “family” variable that you want to compare. Then click “Continue” to go on.

47 Example 2

48 Example 2 Then put the two “question” variables ( , )in as the predictors (“independents”). The sub-dialog box “statistics” lets you see descriptive statistics on each predictor variable for the different groups. Let’s check “Means” to see some basic descriptive statistics.

49 Example 2

50 Example 2

51 Example 2 Click “Continue” to go back to the main dialog box. Then, in the main dialog box, click the “Classify” button to see the next sub-dialog. Here’s the sub-dialog that you get when you click the “Classify” button:

52 Example 2 On this screen, check the box for the “summary table.” This will give you the classification table (sensitivity, specificity, etc.) on your printout. Hit “Continue” to go back to the main dialog box, and then hit “OK” to see the results of your analysis.

53 Example 2 The following is the result. This table just tells you if there’s any missing data.

54 Example 2 This table shows the means that we asked for—it gives means on each variable for people in each sub-group, and also the overall means on each variable.

55 Example 2 Analysis 1 Summary of Canonical Discriminant Functions
This table tells you something about the “latent variable” that you have constructed (the discriminant function), which helps you to differentiate between the groups.

56 Example 2 Here is the multivariate test—Wilks’ lambda.

57 Example 2 These “discriminant function coefficients” work just like the beta-weights in regression. Based on these, you can write out the equation for the discriminant function: Using this equation, given someone’s scores on and , you can calculate their score on the discriminant function. To figure out what that function score means, look at the following group centroids.

58 Example 2 This table tells you the correlation between each item and the discriminant function, but we won’t use it for anything here.

59 Example 2 In practical terms, we usually figure out which group a family is in by calculating a cut score halfway between the two centroids: If an individual family’s score on the function (calculated by plugging in their scores on and to the equation we wrote out above) is above 0, then they were probably lawnmower owners. If their function score is below 0, then they were probably nonowners.

60 Example 2 Classification Statistics

61 Example 2 Here is the classification table that we got by selecting that option in the SPSS dialog box. It gives information about actual group membership and predicted group membership.

62 Summary Discriminant analysis, which is a set of tools and methods used to distinguish between two or more groups of populations and to determine how to allocate new observations into groups. The purpose of discriminant analysis is to estimate the relationship between a single categorical dependent variable and a set of quantitative independent variables. Discriminant analysis plays the same role as the multiple linear regressions by predicting an outcome, and it has widespread applications in situations where the primary objective is identifying the group to which a categorical object belongs.

63 Summary Introduced several well-known discriminant analysis methods, including: Fisher’s discriminant analysis, Bayes discriminant analysis and stepwise discriminant analysis. Show that how to undertake discriminant analysis in SPSS 20.0 through two examples.

64 Exercises 1.What is the purpose of DA? 2.What is the idea of Fisher’s DA? 3.What is stepwise DA? 4.To apply the DA, what conditions should hold?

65 Exercises 5.The following table gives three financial data( , , ) of 10 companies with good financial status (group 1) and 6 companies with poor financial status (group 2). Now a company’s dataset is ( =420.50, =32.42, =1.98). Using DA determine this company is in group 1 or group 2.

66 Exercises 6.The following table gives two weather forecast data ( , )of 8 flood years (group 1) and 8 nonflood years (group 2). Now one year’s dataset is ( =33.6, = -1.5). Using DA determine this year is in group 1 or group 2. Weather forecast data

67 Glossary Discriminant analysis: is a set of methods and tools used to distinguish between two or more groups of populations and to determine how to allocate new observations into groups. Fisher’s discriminant analysis: is to base the discriminant rule on a projection such that a good separation was achieved. Bayes discriminant: is the discriminant involving the posterior distribution. Stepwise discriminant analysis: is a method to find the best set of predictors. Wilks Lambda method: is a method to select predictors that minimize Wilks Lambda.

68 Bibliography [1] Y.T. Zhang and K.T. Fang, Introduction on Multivariate Analysis, Bejing, Science Press, [2] X.Q. He, Multivariate Statistical Analysis (2nd), Beijing, China Renmin University Press, [3] W. Hardle and L. Simar, Applied Multivariate Statistical Analysis, Springer Verlag, Berlin-Heidelberg-New York, [4] G. L. Wang and X.Q. He, Multivariate Economic Data Statistic Analysis, Shaanxi Science Press, [5] N. Johnson and D. Wichren, Applied Multivariate Statistics Analysis, Upper Saddle River, N.J.: Prentice Hall, 1982.

69 End of Chapter 10


Download ppt "Chapter 10 Discriminant Analysis"

Similar presentations


Ads by Google