Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 27 of 42 Friday, 27 October.

Slides:



Advertisements
Similar presentations
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Advertisements

Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
Bayes Rule The product rule gives us two ways to factor a joint probability: Therefore, Why is this useful? –Can get diagnostic probability P(Cavity |
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Bayes Nets Rong Jin. Hidden Markov Model  Inferring from observations (o i ) to hidden variables (q i )  This is a general framework for representing.
Introduction to Bayesian Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán)
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Introduction to Graphical Models.
KI2 - 2 Kunstmatige Intelligentie / RuG Probabilities Revisited AIMA, Chapter 13.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, March 6, 2000 William.
Introduction to Bayesian Learning Ata Kaban School of Computer Science University of Birmingham.
Uncertainty Chapter 13.
Review: Probability Random variables, events Axioms of probability
Computing & Information Sciences Kansas State University Lecture 28 of 42 CIS 530 / 730 Artificial Intelligence Lecture 28 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 26 of 41 Friday, 22 October.
Computing & Information Sciences Kansas State University Lecture 27 of 42 CIS 530 / 730 Artificial Intelligence Lecture 27 of 42 William H. Hsu Department.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Naive Bayes Classifier
Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Lecture 30 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October.
Computing & Information Sciences Kansas State University Wednesday, 15 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 20 of 42 Wednesday, 15 October.
Computing & Information Sciences Kansas State University Monday, 29 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 25 of 42 Wednesday, 29 October.
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Computing & Information Sciences Kansas State University Wednesday, 20 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 12 of 42 Wednesday, 20 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Computing & Information Sciences Kansas State University Lecture 22 of 42 CIS 530 / 730 Artificial Intelligence Lecture 22 of 42 Planning: Sensorless and.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Friday, 29 October 2004 William.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, March 10, 2000 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 of 41 Monday, 25 October.
Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Jim Little Uncertainty 2 Nov 3, 2014 Textbook §6.1.3.
Computing & Information Sciences Kansas State University Wednesday, 25 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 26 of 42 Wednesday. 25 October.
Computing & Information Sciences Kansas State University Lecture 21 of 42 CIS 530 / 730 Artificial Intelligence Lecture 21 of 42 Planning: Graph Planning.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 11 of 41 Wednesday, 15.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 13 of 41 Monday, 20 September.
Computing & Information Sciences Kansas State University Lecture 13 of 42 CIS 530 / 730 Artificial Intelligence Lecture 13 of 42 William H. Hsu Department.
Computing & Information Sciences Kansas State University Lecture 40 of 42 CIS 530 / 730 Artificial Intelligence Lecture 40 of 42 A Brief Survey of Computer.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 17 Wednesday, 01 October.
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Computing & Information Sciences Kansas State University Data Sciences Summer Institute Multimodal Information Access and Synthesis Learning and Reasoning.
Computing & Information Sciences Kansas State University Monday, 06 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 31 of 42 Monday, 06 November.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Chapter 6 Bayesian Learning
Computing & Information Sciences Kansas State University Lecture 14 of 42 CIS 530 / 730 Artificial Intelligence Lecture 14 of 42 William H. Hsu Department.
Computing & Information Sciences Kansas State University Monday, 25 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 14 of 42 Monday, 25 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 41 Wednesday, 22.
Uncertainty Chapter 13. Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Review: Probability Random variables, events Axioms of probability Atomic events Joint and marginal probability distributions Conditional probability distributions.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 21 February 2007.
Computing & Information Sciences Kansas State University Monday, 23 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 25 of 42 Monday, 23 October.
Computing & Information Sciences Kansas State University Friday, 20 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 24 of 42 Friday, 20 October.
Computing & Information Sciences Kansas State University Lecture 12 of 42 CIS 530 / 730 Artificial Intelligence Lecture 12 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 24 of 41 Monday, 18 October.
Bayesian Learning Bayes Theorem MAP, ML hypotheses MAP learners
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 28 February 2007.
Computing & Information Sciences Kansas State University Monday, 09 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 19 of 42 Monday, 09 October.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Computing & Information Sciences Kansas State University Wednesday, 04 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 17 of 42 Wednesday, 04 October.
Computing & Information Sciences Kansas State University Friday, 03 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 30 of 42 Friday, 03 November.
Computing & Information Sciences Kansas State University Friday, 13 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 21 of 42 Friday, 13 October.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Monday, 01 February 2016 William.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
Computing & Information Sciences Kansas State University Wednesday, 25 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 26 of 42 Wednesday. 25 October.
Computing & Information Sciences Kansas State University Monday, 18 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 11 of 42 Monday, 18 September.
Computing & Information Sciences Kansas State University Wednesday, 01 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 29 of 42 Wednesday, 01 November.
Review of Probability.
Naive Bayes Classifier
Reasoning Under Uncertainty: Conditioning, Bayes Rule & Chain Rule
Naive Bayes Classifier
Presentation transcript:

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 27 of 42 Friday, 27 October 2006 William H. Hsu Department of Computing and Information Sciences, KSU KSOL course page: Course web site: Instructor home page: Reading for Next Class: Chapter 13, Russell & Norvig 2 nd edition Uncertain Reasoning: Probability Review Discussion: Probabilistic Reasoning Apps

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture Outline Today’s Reading: Sections 12.5 – 12.8, R&N 2e Next Week’s Reading: Chapter 13, Sections 14.1 – 14.2, R&N 2e Today: Intro to Uncertain Reasoning  Nondeterminism in the real world  Incomplete domain theories  Observation errors: sensor, measurement, estimation  Actuator errors Probability Review  Kolmogorov axioms  Conditioning Next Week: Graphical models  Bayesian networks and causality  Inference and learning  BNJ interface (

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Looking Ahead [1]: Planning and Learning Roadmap Bounded Indeterminacy (12.3) Four Techniques for Dealing with Nondeterministic Domains 1. Sensorless aka Conformant Planning: “Be Prepared” (12.3)  Idea: be able to respond to any situation (universal planning)  Coercion 2. Conditional aka Contingency Planning: “Review the Situation” (12.4)  Idea: be able to respond to many typical alternative situations  Actions for sensing 3. Execution Monitoring and Replanning: “The Show Must Go On” (12.5)  Idea: be able to resume momentarily failed plans  Plan revision 4. Continuous Planning: “Always in Motion, The Future Is” (12.6)  Lifetime planning (and learning!)  Formulate new goals

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Probability: Basic Definitions and Axioms Sample Space (  ): Range of a Random Variable X Probability Measure Pr(  )   denotes a range of “events”; X:   Probability Pr, or P, is a measure over 2   In a general sense, Pr(X = x   ) is a measure of belief in X = x  P(X = x) = 0 or P(X = x) = 1: plain (aka categorical) beliefs (can’t be revised)  All other beliefs are subject to revision Kolmogorov Axioms  1.  x  . 0  P(X = x)  1  2. P(  )   x   P(X = x) = 1  3. Joint Probability: P(X 1  X 2 )  Probability of the Joint Event X 1  X 2 Independence: P(X 1  X 2 ) = P(X 1 )  P(X 2 )

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Basic Formulas for Probabilities Product Rule (Alternative Statement of Bayes’s Theorem)  Proof: requires axiomatic set theory, as does Bayes’s Theorem Sum Rule  Sketch of proof (immediate from axiomatic set theory)  Draw a Venn diagram of two sets denoting events A and B  Let A  B denote the event corresponding to A  B… Theorem of Total Probability  Suppose events A 1, A 2, …, A n are mutually exclusive and exhaustive  Mutually exclusive: i  j  A i  A j =   Exhaustive:  P(A i ) = 1  Then  Proof: follows from product rule and 3 rd Kolmogorov axiom A B

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Bayes’s Theorem [1] Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Looking Ahead [2]: Uncertain Reasoning Roadmap Framework: Interpretations of Probability [Cheeseman, 1985]  Bayesian subjectivist view  A measure of an agent’s belief in a proposition  Proposition denoted by random variable (sample space: range)  e.g., Pr(Outlook = Sunny) = 0.8  Frequentist view: probability is the frequency of observations of an event  Logicist view: probability is inferential evidence in favor of a proposition Some Applications  HCI: learning natural language; intelligent displays; decision support  Approaches: prediction; sensor and data fusion (e.g., bioinformatics) Prediction: Examples  Measure relevant parameters: temperature, barometric pressure, wind speed  Make statement of the form Pr(Tomorrow’s-Weather = Rain) = 0.5  College admissions: Pr(Acceptance)  p  Plain beliefs: unconditional acceptance (p = 1) or categorical rejection (p = 0)  Conditional beliefs: depends on reviewer (use probabilistic model)

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Automated Reasoning using Probabilistic Models: Inference Tasks Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Looking Ahead [3]: Bayesian Network tools in Java (BNJ) © 2004 KSU BNJ Development TeamAsia (Chest Clinic) Network

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence BNJ Core [2] Graph Architecture © 2004 KSU BNJ Development TeamCPCS-54 Network

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Bayes’s Theorem [2] Theorem P(h)  Prior Probability of Assertion (Hypothesis) h  Measures initial beliefs (BK) before any information is obtained (hence prior) P(D)  Prior Probability of Data (Observations) D  Measures probability of obtaining sample D (i.e., expresses D) P(h | D)  Probability of h Given D  | denotes conditioning - hence P(h | D) is a conditional (aka posterior) probability P(D | h)  Probability of D Given h  Measures probability of observing D given that h is correct (“generative” model) P(h  D)  Joint Probability of h and D  Measures probability of observing D and of h being correct

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Bayesian Inference: Query Answering (QA) Answering User Queries  Suppose we want to perform intelligent inferences over a database DB  Scenario 1: DB contains records (instances), some “labeled” with answers  Scenario 2: DB contains probabilities (annotations) over propositions  QA: an application of probabilistic inference QA Using Prior and Conditional Probabilities: Example  Query: Does patient have cancer or not?  Suppose: patient takes a lab test and result comes back positive  Correct + result in only 98% of the cases in which disease is actually present  Correct - result in only 97% of the cases in which disease is not present  Only of the entire population has this cancer    P(false negative for H 0  Cancer) = 0.02 (NB: for 1-point sample)    P(false positive for H 0  Cancer) = 0.03 (NB: for 1-point sample)  P(+ | H 0 ) P(H 0 ) = , P(+ | H A ) P(H A ) =  h MAP = H A   Cancer

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Choosing Hypotheses Bayes’s Theorem MAP Hypothesis  Generally want most probable hypothesis given the training data  Define:  the value of x in the sample space  with the highest f(x)  Maximum a posteriori hypothesis, h MAP ML Hypothesis  Assume that p(h i ) = p(h j ) for all pairs i, j (uniform priors, i.e., P H ~ Uniform)  Can further simplify and choose the maximum likelihood hypothesis, h ML

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Automated Reasoning using Probabilistic Models: Inference Tasks Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Choosing Hypotheses Bayes’s Theorem MAP Hypothesis  Generally want most probable hypothesis given the training data  Define:  the value of x in the sample space  with the highest f(x)  Maximum a posteriori hypothesis, h MAP ML Hypothesis  Assume that p(h i ) = p(h j ) for all pairs i, j (uniform priors, i.e., P H ~ Uniform)  Can further simplify and choose the maximum likelihood hypothesis, h ML

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Graphical Models of Probability P(20s, Female, Low, Non-Smoker, No-Cancer, Negative, Negative) = P(T) · P(F) · P(L | T) · P(N | T, F) · P(N | L, N) · P(N | N) · P(N | N) Conditional Independence  X is conditionally independent (CI) from Y given Z iff P(X | Y, Z) = P(X | Z) for all values of X, Y, and Z  Example: P(Thunder | Rain, Lightning) = P(Thunder | Lightning)  T  R | L Bayesian (Belief) Network  Acyclic directed graph model B = (V, E,  ) representing CI assertions over   Vertices (nodes) V: denote events (each a random variable)  Edges (arcs, links) E: denote conditional dependencies Markov Condition for BBNs (Chain Rule): Example BBN X1X1 X3X3 X4X4 X5X5 Age Exposure-To-Toxins Smoking Cancer X6X6 Serum Calcium X2X2 Gender X7X7 Lung Tumor

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Semantics of Bayesian Networks Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Markov Blanket Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Constructing Bayesian Networks: The Chain Rule of Inference Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Example: Evidential Reasoning for Car Diagnosis Adapted from slides by S. Russell, UC Berkeley

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence BNJ Visualization [2] Pseudo-Code Annotation (Code Page) © 2004 KSU BNJ Development Team ALARM Network

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence BNJ Visualization [3] Network © 2004 KSU BNJ Development Team Poker Network

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Terminology Introduction to Reasoning under Uncertainty  Probability foundations  Definitions: subjectivist, frequentist, logicist  (3) Kolmogorov axioms Bayes’s Theorem  Prior probability of an event  Joint probability of an event  Conditional (posterior) probability of an event Maximum A Posteriori (MAP) and Maximum Likelihood (ML) Hypotheses  MAP hypothesis: highest conditional probability given observations (data)  ML: highest likelihood of generating the observed data  ML estimation (MLE): estimating parameters to find ML hypothesis Bayesian Inference: Computing Conditional Probabilities (CPs) in A Model Bayesian Learning: Searching Model (Hypothesis) Space using CPs

Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Summary Points Introduction to Probabilistic Reasoning  Framework: using probabilistic criteria to search H  Probability foundations  Definitions: subjectivist, objectivist; Bayesian, frequentist, logicist  Kolmogorov axioms Bayes’s Theorem  Definition of conditional (posterior) probability  Product rule Maximum A Posteriori (MAP) and Maximum Likelihood (ML) Hypotheses  Bayes’s Rule and MAP  Uniform priors: allow use of MLE to generate MAP hypotheses  Relation to version spaces, candidate elimination Next Week: Chapter 14, Russell and Norvig  Later: Bayesian learning: MDL, BOC, Gibbs, Simple (Naïve) Bayes  Categorizing text and documents, other applications