Presentation is loading. Please wait.

Presentation is loading. Please wait.

Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.

Similar presentations


Presentation on theme: "Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models."— Presentation transcript:

1 Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models

2 Why probabilistic methods? Probabilistic is used for description of events or behavior or making decision when we have not enough knowledge about object of observation Probability theory and then probabilistic methods of AI aim to introduce in random processes any knowledge about laws or rules about sources of random events This is alternative path of description of uncertainty in contrast to fuzzy logics

3 The Problem We normally deal with assertions and their causal connections: –John has fever –John has the flu –If somebody has the flu then that person has fever. We are not certain that such assertions are true. We believe/disbelieve them to some degree. [Though "belief" and "evidence" are not the same thing, for our purposes they will be treated synonymously.] Our problem is how to associate a degree of belief or of disbelief with assertions –How do we associate beliefs with elementary assertions –How do we combine beliefs in composite assertions from the beliefs of the component assertions –What is the relation between the beliefs of causally connected assertions. Estimates for elementary assertions are obtained –From Experts (subjective probability) –From frequencies (if given enough data) It is very hard to come up with good estimates for beliefs. Always consider the question "What if the guess is bad". Estimates are needed, given the belief in assertions A and B, for the assertions ~A, A & B, A v B Evidence must be combined in cases such as: –We have a causal connection from assertion A to assertion B, what can we say about B if A is true, or, vice versa, about A if B is true –We have a causal connection from assertion A to assertions B1 and B2, what can we say about A if both B1 and B2 are true –We have a causal connection from assertion A1 to B and a causal connection from A2 to B, what can we say about B when both A1 and A2 are true.

4 Probabilistic methods of reasoning and learning Probabilistic neural networks Bayesian networks Markov models and chains Support Vector and Kernel Machines (SVM) Genetic algorithms (evolution learning)

5 Bayes’ Law – P(a,b) = P(a|b) P(b) = P(b|a) P(a) –Joint probability of a and b = probability of b times the probability of a given b

6 Bayesian learning

7 Bayesian learning (2)

8 Bayesian learning (3)

9 Bayesian learning (4)

10 Bayes theorem P(A j | B) – posterior probability of event A j at condition of event B, P(B | A j ) – likelihood, P(B) – evidence Bayes theorem is only valid if we know all the conditional probabilities relating to the evidence in question. This makes it hard to apply the theorem in practical AI applications

11 Bayesian Network A Bayesian Network is a directed acyclic graph: –A graph where the directions are links which indicate dependencies that exist between nodes (variables). –Nodes represent propositions about events or events themselves. –Conditional probabilities quantify the strength of dependencies. Consider the following example: The probability, that my car won't start. If my car won't start then it is likely that –The battery is flat or –The staring motor is broken. In order to decide whether to fix the car myself or send it to the garage I make the following decision: –If the headlights do not work then the battery is likely to be flat so i fix it myself. –If the starting motor is defective then send car to garage. –If battery and starting motor both gone send car to garage.

12 A simple Bayesian network

13 Kinds of relations between variables in Bayesian nets a) Sequence, influence may be distribute from A to C and back while value of B is unknown b) Divergence, influence may be distributed on childes of A while A is unknown c) Convergence, about A nothing unknown except that may be obtained through its parents

14 Reasoning in Bayesian nets Probabilities in links obey standard conditional probability axioms. Therefore follow links in reaching hypothesis and update beliefs accordingly. A few broad classes of algorithms have bee used to help with this: –Pearls's message passing method. –Clique triangulation. –Stochastic methods. –Basically they all take advantage of clusters in the network and use their limits on the influence to constrain the search through net. –They also ensure that probabilities are updated correctly. Since information is local information can be readily added and deleted with minimum effect on the whole network. ONLY affected nodes need updating.

15 Synthesis of Bayes network based on a priory information Describe task in terms of probabilities of values of goal variables Select concept space of task, determine variables corresponding to goal variables, describe possible values of ones Determine a priori probabilities of values of variables Describe causal relations and node (variables) as graph For every node determine condition probabilities of value of variable at different combinations of values of variables-parents

16 Applications of Bayes networks Medical diagnostic systems –PathFinder (1990) for diagnostics of illness of lymphatic glands, Space and military applications –Vista (NASA) is used for selection of needed information for diagnostic display from telemetric information in real time, –Netica (Australia) for defence of territory from sea Computers and software –For control of agents-helpers in MS Office Image processing –Extract of 3-dimensional scene from 2-dimensional images Finance and economy –Estimation of risks and prediction of yield of portfolio

17

18

19

20

21 Hidden Markov Model for recognition of speech P3(.)P3(.)P1(.)P1(.)P2(.)P2(.) P 1,1 P 1,2 P 2,2 P 3,3 P 2,3 P 3,4

22 Create compound HMM for each lexical entry by concatenating the phones making up the pronunciation –example of HMM for ‘lab’ (following ‘speech’ for crossword triphone) Multiple pronunciations can be weighted by likelihood into compound HMM for a word (Tri)phone models are independent parts of word models Lexical HMMs P3(.)P3(.)P1(.)P1(.)P2(.)P2(.) P 1,1 P 1,2 P 2,2 P 3,3 P 2,3 P 3,4 P3(.)P3(.)P1(.)P1(.)P2(.)P2(.) P 1,1 P 1,2 P 2,2 P 3,3 P 2,3 P 3,4 P3(.)P3(.)P1(.)P1(.)P2(.)P2(.) P 1,1 P 1,2 P 2,2 P 3,3 P 2,3 P 3,4 phone: l a b triphone: ch- l +a l- a +b a- b +#


Download ppt "Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models."

Similar presentations


Ads by Google