Download presentation

Presentation is loading. Please wait.

Published byAustin Maddux Modified over 3 years ago

1
Data Mining and Machine Learning Naive Bayes David Corne, HWU dwcorne@gmail.com

2
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY

3
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY A new patient has a blood test – his P34 level is HIGH. what is our best guess for prostate cancer?

4
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY It’s useful to know: P(cancer = Y)

5
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY It’s useful to know: P(cancer = Y) - on basis of this tiny dataset, P(c = Y) is 5/10 = 0.5

6
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY But we know that P34 =H, so actually we want: P(cancer=Y | P34 = H) - the prob that cancer is Y, given that P34 is high

7
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY P(cancer=Y | P34 = H) - the prob that cancer is Y, given that P34 is high - this seems to be 2/3 = ~ 0.67

8
A very simple dataset – one field / one class P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY So we have: P ( c=Y | P34 = H) = 0.67 P ( c =N | P34 = H) = 0.33 The class value with the highest probability is our best guess

9
In general we may have any number of class values P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N Maybe MediumY suppose again we know that P34 is High; here we have: P ( c=Y | P34 = H) = 0.5 P ( c=N | P34 = H) = 0.25 P(c = Maybe | H) = 0.25... and again, Y is the winner

10
That is the essence of Naive Bayes, but: the probability calculations are much trickier when there are >1 fields so we make a ‘Naive’ assumption that makes it simpler

11
Bayes’ theorem P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY As we saw, on the right we are illustrating: P(cancer = Y | P34 = H)

12
Bayes’ theorem P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY And now we are illustrating P(P34 = H | cancer = Y) This is a different thing, that turns out as 2/5 = 0.4

13
Bayes’ theorem is this: P( A | B) = P ( B | A ) P (A) P(B) It is very useful when it is hard to get P(A | B) directly, but easier to get the things on the right

14
Bayes’ theorem in 1-non-class-field DMML context: P( Class=X | Fieldval = F) = P ( Fieldval = F | Class = X ) × P( Class = X) P(Fieldval = F)

15
Bayes’ theorem in 1-non-class-field DMML context: P( Class=X | Fieldval = F) = P ( Fieldval = F | Class = X ) × P( Class = X) P(Fieldval = F) We want to check this for each class and choose the class that gives the highest value.

16
Bayes’ theorem in 1-non-class-field DMML context: P( Class=X | Fieldval = F) = P ( Fieldval = F | Class = X ) × P( Class = X) P(Fieldval = F) E.g. We compare: P(F | Yes) × P (Yes) P(F | No× P (No) P(F | Maybe) × P (Maybe)... We can ignore the P(Fieldval = F) bit... Why ?

17
and that was Exactly how we do Naive Bayes for a 1-field dataset

18
Note how this relates to our beloved histograms P34 levelProstate cancer HighY MediumY LowY N N MediumN HighY N LowN MediumY Low Med High 0.6 0.4 0.2 0 P(L| N) P(H| Y)

19
Nave-Bayes with Many-fields P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y

20
Nave-Bayes with Many-fields P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ?

21
Nave-Bayes with Many-fields P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ? P(p34=M | Y) × P(p61=M | Y) × P(BMI=H |Y) × P(cancer = Y) P(p34=M | N) × P(p61=M | N) × P(BMI=H |N) × P(cancer = N) which of these gives the highest value?

22
Nave-Bayes with P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ? P(p34=M | Y) × P(p61=M | Y) × P(BMI=H |Y) × P(cancer = Y) P(p34=M | N) × P(p61=M | N) × P(BMI=H |N) × P(cancer = N) which of these gives the highest value?

23
Nave-Bayes with Many-fields P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ? P(p34=M | Y) × P(p61=M | Y) × P(BMI=H |Y) × P(cancer = Y) P(p34=M | N) × P(p61=M | N) × P(BMI=H |N) × P(cancer = N) which of these gives the highest value?

24
Nave-Bayes with Many-fields P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ? P(p34=M | Y) × P(p61=M | Y) × P(BMI=H |Y) × P(cancer = Y) P(p34=M | N) × P(p61=M | N) × P(BMI=H |N) × P(cancer = N) which of these gives the highest value?

25
Nave-Bayes with P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ? P(p34=M | Y) × P(p61=M | Y) × P(BMI=H |Y) × P(cancer = Y) P(p34=M | N) × P(p61=M | N) × P(BMI=H |N) × P(cancer = N) which of these gives the highest value?

26
Nave-Bayes with Many-fields P34 levelP61 levelBMIProstate cancer HighLowMediumY LowMediumY Low HighY LowHighLowN N Medium LowN HighLowMediumY HighMediumLowN HighN MediumHigh Y New patient: P34=M, P61=M, BMI = H Best guess at cancer field ? 0.4 × 0 × 0.4 × 0.5 = 0 0.2 × 0.4 × 0.2 × 0.5 = 0.008 which of these gives the highest value?

27
Nave-Bayes -- in general N fields, q possible class values, New unclassified instance: F1 = v1, F2 = v2,..., Fn = vn what is the class value? i.e. Is it c1, c2,.. or cq ? calculate each of these q things – biggest one gives the class: P(F1=v1 | c1) × P(F2=v2 | c1) ×... × P(Fn=vn | c1) × P(c1) P(F1=v1 | c2) × P(F2=v2 | c2) ×... × P(Fn=vn | c2) × P(c2)... P(F1=v1 | cq) × P(F2=v2 | cq) ×... × P(Fn=vn | cq) × P(cq)

28
Nave-Bayes -- in general Actually... What we normally do, when there are more than a handful of fields, is this Calculate: log(P(F1=v1 | c1)) +... + log(P(Fn=vn | c1)) + log( P(c1)) log(P(F1=v1 | c2)) +... + log(P(Fn=vn | c2)) + log( P(c2)) and choose class based on highest of these. Why ?

29
Nave-Bayes -- in general Because log( a × b × c × …) = log(a) + log(b) + log(c) + … and this means we won’t get underflow errors, which we would otherwise get with, e.g. 0.003 × 0.000296 × 0.001 × …[100 fields] × 0.042 …

30
Deriving NB Essence of Naive Bayes, with 1 non-class field, is to calc this for each class value, given some new instance with fieldval = F: P(class = C | Fieldval = F) For many fields, our new instance is (e.g.) (F1, F2,...Fn), and the ‘essence of Naive Bayes’ is to calculate this for each class: P(class = C | F1,F2,F3,...,Fn) i.e. What is prob of class C, given all these field vals together?

31
Apply magic dust and Bayes theorem, and...... If we make the naive assumption that all of the fields are independent of each other (e.g. P(F1| F2) = P(F1), etc...)... then P (class = C | F1,F2,F3,...,Fn) = P( F1 and F2 and... and Fn | C) x P (C) = P(F1| C) x P (F2 | C) x... X P(Fn | C) x P(C) Which is what we calculate in NB

32
EN(B)D

Similar presentations

OK

Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.

Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.

© 2018 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on computer languages wikipedia Ppt on statistics in maths what does median Ppt on bank management system in java Ppt on total internal reflection definition Ppt on girl child death in india Ppt on tcp/ip protocol and osi model Ppt on establishment of company power in india Download ppt on orthogonal frequency division multiplexing Ppt on resistance temperature detector rtd Ppt on regional trade agreements notified