Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 2-Concept Learning (1/3) Eduardo Poggi

Similar presentations


Presentation on theme: "1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 2-Concept Learning (1/3) Eduardo Poggi"— Presentation transcript:

1 1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 2-Concept Learning (1/3) Eduardo Poggi (eduardopoggi@yahoo.com.ar)eduardopoggi@yahoo.com.ar Ernesto Mislej (emislej@gmail.com)emislej@gmail.com otoño de 2008

2 2 Agenda Definitions Search Space and General-Specific Ordering Concept learning as search FIND-S

3 3 Definition The problem is to learn a function mapping examples into two classes: positive and negative. We are given a database of examples already classified as positive or negative. Concept learning: the process of inducing a function mapping input examples into a Boolean output. Examples: Classifying objects in astronomical images as stars or galaxies Classifying animals as vertebrates or invertebrates

4 4 Working Example: Mushrooms Class of Tasks: Predicting poisonous mushrooms Performance: Accuracy of Classification Experience: Database describing mushrooms with their class Knowledge to learn: Function mapping mushrooms to {+,-} where -:not-poisonous and +:poisonous where -:not-poisonous and +:poisonous Representation of target knowledge: conjunction of attribute values. Learning mechanism: candidate-elimination

5 5 Notation Set of instances X Target concept c : X  {+,-} Training examples E = {(x, c(x))} Data set D  X Set of possible hypotheses H h  H h : X  {+,-} Goal:Find h / h(x)=c(x)

6 6 Representation of Examples Features: color {red, brown, gray} color {red, brown, gray} size {small, large} size {small, large} shape {round,elongated} shape {round,elongated} land {humid,dry} land {humid,dry} air humidity {low,high} air humidity {low,high} texture {smooth, rough} texture {smooth, rough}

7 7 The Input and Output Space X Only a small subset is contained in our database. Y = {+,-} X : The space of all possible examples (input space). Y: The space of classes (output space). An example in X is a feature vector X. For instance: X = (red,small,elongated,humid,low,rough) X is the cross product of all feature values. X is the cross product of all feature values.

8 8 The Training Examples D : The set of training examples. D is a set of pairs { (x,c(x)) }, where c is the target concept Example of D: ((red,small,round,humid,low,smooth), +) ((red,small,elongated,humid,low,smooth),+) ((gray,large,elongated,humid,low,rough), -) ((red,small,elongated,humid,high,rough), +) Instances from the input space Instances from the output space

9 9 Hypothesis Representation Consider the following hypotheses: (*,*,*,*,*,*): all mushrooms are poisonous (*,*,*,*,*,*): all mushrooms are poisonous (0,0,0,0,0,0): no mushroom is poisonous (0,0,0,0,0,0): no mushroom is poisonous Special symbols:  * Any value is acceptable  0 no value is acceptable Any hypothesis h is a function from X to Y h: X Y h: X Y We will explore the space of conjunctions.

10 10 Hypothesis Space The space of all hypotheses is represented by H The space of all hypotheses is represented by H Let h be a hypothesis in H. Let h be a hypothesis in H. Let X be an example of a mushroom. Let X be an example of a mushroom. if h(X) = + then X is poisonous, if h(X) = + then X is poisonous, otherwise X is not-poisonous Our goal is to find the hypothesis, h*, that is very “close” Our goal is to find the hypothesis, h*, that is very “close” to target concept c. to target concept c. A hypothesis is said to “cover” those examples it classifies A hypothesis is said to “cover” those examples it classifies as positive. as positive. X h

11 11 Assumption 1 We will explore the space of all conjunctions. We assume the target concept falls within this space. Target concept c H

12 12 Assumption 2 A hypothesis close to target concept c obtained after seeing many training examples will result in high accuracy on the set of unobserved examples. Training set D Hypothesis h* is good Complement set D’ Hypothesis h* is good

13 13 Concept Learning as Search There is a general to specific ordering inherent to any hypothesis space. Consider these two hypotheses: h1 = (red,*,*,humid,*,*) h2 = (red,*,*,*,*,*) We say h2 is more general than h1 because h2 classifies more instances than h1 and h1 is covered by h2.

14 14 General-Specific For example, consider the following hypotheses: h1 h2h3 h1 is more general than h2 and h3. h2 and h3 are neither more specific nor more general than each other.

15 15 Let hj and hk be two hypotheses mapping examples into {+,-}. We say hj is more general than hk iff For all examples X, hk(X) = +  hj(X) = + We represent this fact as hj >= hk The >= relation imposes a partial ordering over the hypothesis space H (reflexive, antisymmetric, and transitive). Definition

16 16 Lattice Any input space X defines then a lattice of hypotheses ordered according to the general-specific relation: h1 h3h4 h2 h5h6 h7 h8

17 17 Finding a Maximally-Specific Hypothesis Algorithm to search the space of conjunctions:  Start with the most specific hypothesis  Generalize the hypothesis when it fails to cover a positive example Algorithm: 1.Initialize h to the most specific hypothesis 2.For each positive training example X For each value a in h For each value a in h If example X and h agree on a, do nothing If example X and h agree on a, do nothing else generalize a by the next more general constraint else generalize a by the next more general constraint 3. Output hypothesis h

18 18 Example Let’s run the learning algorithm above with the following examples: ((red,small,round,humid,low,smooth), +) ((red,small,elongated,humid,low,smooth),+) ((gray,large,elongated,humid,low,rough), -) ((red,small,elongated,humid,high,rough), +) We start with the most specific hypothesis: h = (0,0,0,0,0,0) The first example comes and since the example is positive and h fails to cover it, we simply generalize h to cover exactly this example: h = (red,small,round,humid,low,smooth)

19 19 Example Hypothesis h basically says that the first example is the only positive example, all other examples are negative. Then comes examples 2: ((red,small,elongated,humid,low,smooth), poisonous) This example is positive. All attributes match hypothesis h except for attribute shape: it has the value elongated, not round. We generalize this attribute using symbol * yielding: h: (red,small,*,humid,low,smooth) The third example is negative and so we just ignore it. Why is it we don’t need to be concerned with negative examples?

20 20 Example Upon observing the 4 th example, hypothesis h is generalized to the following: h = (red,small,*,humid,*,*) h is interpreted as any mushroom that is red, small and found on humid land should be classified as poisonous.

21 21 Analyzing the Algorithm The algorithm is guaranteed to find the hypothesis that is most specific and consistent with the set of training examples.The algorithm is guaranteed to find the hypothesis that is most specific and consistent with the set of training examples. It takes advantage of the general-specific ordering to move on the corresponding lattice searching for the next most specific hypothesis.It takes advantage of the general-specific ordering to move on the corresponding lattice searching for the next most specific hypothesis. h1 h3h4 h2 h5h6 h7 h8

22 22 X-H Relation

23 23 X-H Relation

24 24 Points to Consider There are many hypotheses consistent with the training data D. There are many hypotheses consistent with the training data D. Why should we prefer the most specific hypothesis? Why should we prefer the most specific hypothesis? What would happen if the examples are not consistent? What would happen if the examples are not consistent? What would happen if they have errors, noise? What would happen if they have errors, noise? What if there is a hypothesis space H where one can find more that one maximally specific hypothesis h? What if there is a hypothesis space H where one can find more that one maximally specific hypothesis h? The search over the lattice must then be different to allow for this possibility. The search over the lattice must then be different to allow for this possibility.

25 25 Summary The input space is the space of all examples; the output space is the space of all classes. The input space is the space of all examples; the output space is the space of all classes. A hypothesis maps examples into classes. A hypothesis maps examples into classes. We want a hypothesis close to target concept c. We want a hypothesis close to target concept c. The input space establishes a partial ordering over the hypothesis space. The input space establishes a partial ordering over the hypothesis space. One can exploit this ordering to move along the corresponding lattice. One can exploit this ordering to move along the corresponding lattice.

26 26 Tareas Leer Capítulo 2 de Mitchell (-2.5)


Download ppt "1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 2-Concept Learning (1/3) Eduardo Poggi"

Similar presentations


Ads by Google