1 Reasoning Under Uncertainty Artificial Intelligence Chapter 9.

Slides:



Advertisements
Similar presentations
FT228/4 Knowledge Based Decision Support Systems
Advertisements

Lahore University of Management Sciences, Lahore, Pakistan Dr. M.M. Awais- Computer Science Department 1 Lecture 12 Dealing With Uncertainty Probabilistic.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
FT228/4 Knowledge Based Decision Support Systems
INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio LECTURE 11 (Lab): Probability reminder.
5/17/20151 Probabilistic Reasoning CIS 479/579 Bruce R. Maxim UM-Dearborn.
AI – CS364 Uncertainty Management Introduction to Uncertainty Management 21 st September 2006 Dr Bogdan L. Vrusias
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
AI – CS364 Uncertainty Management 26 th September 2006 Dr Bogdan L. Vrusias
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Chapter 4 Probability.
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 12 Jim Martin.
KI2 - 2 Kunstmatige Intelligentie / RuG Probabilities Revisited AIMA, Chapter 13.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Representing Uncertainty CSE 473. © Daniel S. Weld 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one.
Ai in game programming it university of copenhagen Welcome to... the Crash Course Probability Theory Marco Loog.
CSE (c) S. Tanimoto, 2008 Bayes Nets 1 Probabilistic Reasoning With Bayes’ Rule Outline: Motivation Generalizing Modus Ponens Bayes’ Rule Applying.
I The meaning of chance Axiomatization. E Plurbus Unum.
Lecture 05 Rule-based Uncertain Reasoning
CSc411Artificial Intelligence1 Chapter 5 STOCHASTIC METHODS Contents The Elements of Counting Elements of Probability Theory Applications of the Stochastic.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Uncertainty Chapter 13.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 14 From Randomness to Probability.
For Monday after Spring Break Read Homework: –Chapter 13, exercise 6 and 8 May be done in pairs.
Methods in Computational Linguistics II Queens College Lecture 2: Counting Things.
CSCI 121 Special Topics: Bayesian Network Lecture #1: Reasoning Under Uncertainty.
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Chapter 11 Chi-Square Procedures 11.3 Chi-Square Test for Independence; Homogeneity of Proportions.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
1 Chapter 13 Uncertainty. 2 Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Artificial Intelligence CS 165A Tuesday, November 20, 2007  Knowledge Representation (Ch 10)  Uncertainty (Ch 13)
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
CSE PR 1 Reasoning - Rule-based and Probabilistic Representing relations with predicate logic Limitations of predicate logic Representing relations.
Uncertainty Uncertain Knowledge Probability Review Bayes’ Theorem Summary.
Uncertainty Management in Rule-based Expert Systems
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Uncertainty in Expert Systems
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Uncertainty Chapter 13. Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Copyright © 2010 Pearson Education, Inc. Unit 4 Chapter 14 From Randomness to Probability.
4 Proposed Research Projects SmartHome – Encouraging patients with mild cognitive disabilities to use digital memory notebook for activities of daily living.
Uncertainty Chapter 13. Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Reasoning with Uncertainty دكترمحسن كاهاني
Bayes Theorem. Prior Probabilities On way to party, you ask “Has Karl already had too many beers?” Your prior probabilities are 20% yes, 80% no.
Textbook Basics of an Expert System: – “Expert systems: Design and Development,” by: John Durkin, 1994, Chapters 1-4. Uncertainty (Probability, Certainty.
Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? Problems: 1.partial observability (road state, other.
Chapter 14 From Randomness to Probability. Dealing with Random Phenomena A random phenomenon: if we know what outcomes could happen, but not which particular.
Probabilistic Robotics Introduction Probabilities Bayes rule Bayes filters.
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
CSE (c) S. Tanimoto, 2007 Bayes Nets 1 Bayes Networks Outline: Why Bayes Nets? Review of Bayes’ Rule Combining independent items of evidence General.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
AP Statistics From Randomness to Probability Chapter 14.
. Chapter 14 From Randomness to Probability. Slide Dealing with Random Phenomena A is a situation in which we know what outcomes could happen, but.
CHAPTER 5 Handling Uncertainty BIC 3337 EXPERT SYSTEM.
Chapter 10: Using Uncertain Knowledge
Uncertainty Chapter 13.
Basic Probabilistic Reasoning
Professor Marie desJardins,
CS 188: Artificial Intelligence Fall 2008
Honors Statistics From Randomness to Probability
Class #21 – Monday, November 10
Probabilistic Reasoning With Bayes’ Rule
28th September 2005 Dr Bogdan L. Vrusias
Probabilistic Reasoning With Bayes’ Rule
Probabilistic Reasoning With Bayes’ Rule
basic probability and bayes' rule
Presentation transcript:

1 Reasoning Under Uncertainty Artificial Intelligence Chapter 9

2 Probability Part 1

3 Uncertainty Hard for an agent to handle Usually handled statistically Statistical reasoning requires knowledge of probability

4 Uncertainty We want to build reasoning systems that deal with uncertainty by using the laws of probability theory to enable them to reach rational decisions even when there is not enough information to know for certain what the right decision should be.

5 Uncertainty We can’t usually handle uncertainty by means of predicate logic: Computational complexity: Listing all the antecedents and consequences in the agent's environment is too big a job Lack of domain knowledge: We can’t list every possibility because we do not know enough about the domain Pragmatic factors: Use of predicate logic may be impractical for other reasons

6 Probability Probability deals with uncertainty by assigning a numerical value to something Range of values: real number between 0 and 1 Sometime thought of as a bet - what are the odds? Typically concerned with: – Evidence, or probability of observing some particular piece of evidence – Hypothesis, or probability that some hypothesis is true

7 Probability Probability is just a number indicating our belief in a statement. It is often taken to be our degree of willingness to bet on a particular outcome. It varies between 0 (no belief) and 1(absolute confidence).

8 Axioms of probability For any events A, B – 0 ≤ P(A) ≤ 1 – P(Ω) = 1 and P(φ) = 0 – P(A ∨ B) = P(A) + P(B) - P(A ∧ B)

9 Prior Probability P(H) is the a priori probability that a specified hypothesis is true. This is often called the prior probability, or just the prior. This is the unconditional probability, without taking any evidence into consideration.

10 Prior Probability Example: Assume that public health experts have examined 1000,000 people at random, and found that 10 of them have measles. Now we pick a person at random; what is the a priori probability - without any evidence of any kind - that the person has the measles?

11 Prior Probability P(H) = , where H = “has the measles” The a priori probability of someone having the measles is

12 Prior Probability Note that probability only measures our degree of belief in a given proposition. We use probability due to our uncertainty, our lack of knowledge. A particular person always either has the measles or doesn’t. If we knew, we wouldn’t use probability.

13 Prior Probability What is P(  H)? That is, what is the a priori probability of someone not having the measles? P(  H) = 1 - P(H) P(  H) = = 0.999

14 Posterior Probability P(H|E) is the probability that hypothesis H is true given that evidence E is observed. We sometimes read this as, “ the probability of H being true in light of evidence E”. P(H|E) is called the a posteriori probability, or the posterior probability, of H.

15 Posterior Probability Example: Suppose we see red spots (E) on a person’s body, and we consider the hypothesis that they have the measles (H). Does the presence of the evidence change the probability that they have the measles?

16 Posterior Probability If the evidence is relevant to the hypothesis, then yes, we would expect it to change the probability that the hypothesis is true. It’s just common sense that if you observe someone who has red spots on the body, the probability of that person having measles is greater than if you just picked someone at random.

17 Posterior Probability Red spots don’t always indicate measles (they could be due to chicken pox, or something else). Medical surveys show that: P(H|E) = 0.6 The posterior probability that hypothesis H (has measles) is true given that evidence E (red spots) is observed is 0.6.

18 Posterior Probability When would evidence not increase the probability that a hypothesis is true? When the probability of the hypothesis is already 1.0. When the evidence is irrelevant to the hypothesis.

19 Conditional Probability P(E|H) is the probability that evidence E will be observed assuming that hypothesis H is true. We sometimes read this as, “ the probability that hypothesis H being true will result in evidence E”.

20 Conditional Probability Example: Assume that a person has the measles; what is the probability that he will exhibit red spots on the body?

21 Conditional Probability Most people who have the measles eventually get red spots on the body, but not everyone does, and the red spots take a little while to show up after you get the measles. Medical studies show that: P(E|H) =.80 The probability that you will observe a specific piece of evidence (red spots), given a particular hypothesis (that you have the measles), is 0.80.

22 Conditional Probability P(E|  H) is the probability that evidence E will be observed even if hypothesis H is not true. What is the probability that someone will have red spots if they don’t have the measles? Someone who doesn’t have the measles may still have red spots (due to chicken pox, etc.). The experts tell us that: P(E|  H) = 0.10

23 Joint Probability What are all the possible combinations of symptoms? 2 symptoms: fever and red spots 3 symptoms: fever, red spots, malaise 4 symptoms: fever, red spots, malaise, depressed white blood cell count, etc. In general, it is not practical to define all 2 n entries for the n Boolean variables - the full joint probability.

24 Joint Probability P (H  E) is the probability that both H is true and E is observed. This is called the joint probability of H and E.

25 Joint Probability We can express the conditional probability, P(H|E) (probability that hypothesis H is true given that evidence E is observed) in terms of the joint probability P(H  E), (probability that both H is true and E is observed).

26 Joint Probability P(H|E) =P(H  E) P(E) This states that the probability that hypothesis H is true if evidence E is observed is equal to the joint probability of H and E, divided by the prior probability of E. Suppose that we observe 10 people who have red spots (E), and, out of those, 6 have measles (H  E). Then: P(H|E) = 0.6

27 Joint Probability P(H|E) =P(H  E) P(E) and similarly P(E|H) =P(E  H) P(H) Multiplying by P(H) we get P(E  H) = P(E|H) P(H)

28 Product rule Product Rule: P(E  H) = P(E|H)P(H) = P(H|E)P(E) The probability of A and B being true is equal to the probability of A being true if we observe B, times the probability that B is true. The probability of A and B being true is equal to the probability of B being true if we observe A, times the probability that A is true.

29 Deriving Bayes’s Rule P(E  H) = P(E|H) P(H) (from before) Since joint probability is commutative, we have P(H  E) = P(E  H) and therefore P(H  E) = P(E|H) P(H) Substituting that into P(H|E) =P(H  E) P(E) gives us P(H|E) =P(E|H) P(H) (Bayes’s rule) P(E)

30 Conditional Probability Bayes' Rule: P(H|E) =P(E|H) P(H) P(E) This says that the probability that the hypothesis H is true, given that we observe some evidence E, is equal to the probability that we will observe some evidence E if the hypothesis H is true, times the a priori probability that H is true, all divided by the a priori probability that we will observe evidence E.

31 Bayes’s Rule P(E) = P(E|H)P(H) + P(E|  H)P(  H) The a priori probability of observing some evidence E is equal to the probability that we will observe some evidence E if the hypothesis H is true, times the a priori probability that H is true, plus the probability of observing E if the hypothesis is not true, times the a priori probability that H is not true.

32 Conditional Probability Bayes‘s Rule: And this can be generalized to cover situations where there are multiple possible hypotheses.

33 Why use Bayes’s Rule? Knowledge of causes: P (red spots | measles) is often easier to obtain than diagnostic knowledge: P (measles | red spots). Bayes’s Rule lets us use knowledge of causes to derive diagnostic knowledge.

34 Conditional independence But often we must compute the full joint probability of many competing hypotheses and many symptoms - computationally expensive. Instead, we make the assumption of conditional independence. X and Y are conditionally independent given Z iff: P(X,Y | Z) = P(X|Z) P(Y|Z)

35 Conditional independence Assuming conditional independence allows us to avoid computing the full joint probability. Assuming conditional independence is more often justifiable than assuming complete independence.

36 Conditional independence Assume red spots and fever are conditionally independent given measles: P(red spots, fever | measles) = P(red spots | measles) P (fever | measles) Then we can deduce: P(red spots, fever, measles) = P(red spots, fever | measles) P(measles) = P(red spots | measles) P(fever | measles) P(measles)