Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October.

Slides:



Advertisements
Similar presentations
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Advertisements

ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Probability: Review The state of the world is described using random variables Probabilities are defined over events –Sets of world states characterized.
Lecture 5 Bayesian Learning
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
IMPORTANCE SAMPLING ALGORITHM FOR BAYESIAN NETWORKS
Probabilistic inference
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Introduction to Bayesian Learning Bob Durrant School of Computer Science University of Birmingham (Slides: Dr Ata Kabán)
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, March 6, 2000 William.
Introduction to Bayesian Learning Ata Kaban School of Computer Science University of Birmingham.
Bayes Classification.
Jeff Howbert Introduction to Machine Learning Winter Classification Bayesian Classifiers.
Review: Probability Random variables, events Axioms of probability
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Tuesday 15 October 2002 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 26 of 41 Friday, 22 October.
Naive Bayes Classifier
Computing & Information Sciences Kansas State University Lecture 30 of 42 CIS 530 / 730 Artificial Intelligence Lecture 30 of 42 William H. Hsu Department.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 21 of 41 Wednesday, 08.
Computing & Information Sciences Kansas State University Monday, 29 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 25 of 42 Wednesday, 29 October.
CS464 Introduction to Machine Learning1 Bayesian Learning Features of Bayesian learning methods: Each observed training example can incrementally decrease.
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Friday, 29 October 2004 William.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, March 10, 2000 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 of 41 Monday, 25 October.
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Computing & Information Sciences Kansas State University Wednesday, 25 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 26 of 42 Wednesday. 25 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 11 of 41 Wednesday, 15.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 13 of 41 Monday, 20 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 17 Wednesday, 01 October.
Computing & Information Sciences Kansas State University Monday, 06 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 31 of 42 Monday, 06 November.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
Chapter 6 Bayesian Learning
Computing & Information Sciences Kansas State University Monday, 25 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 14 of 42 Monday, 25 September.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 41 Wednesday, 22.
Review: Probability Random variables, events Axioms of probability Atomic events Joint and marginal probability distributions Conditional probability distributions.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 15 of 41 Friday 24 September.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 21 February 2007.
Computing & Information Sciences Kansas State University Friday, 27 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 27 of 42 Friday, 27 October.
Computing & Information Sciences Kansas State University Monday, 23 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 25 of 42 Monday, 23 October.
Computing & Information Sciences Kansas State University Friday, 20 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 24 of 42 Friday, 20 October.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Thursday, 08 March 2007 William.
Kansas State University Department of Computing and Information Sciences CIS 798: Intelligent Systems and Machine Learning Tuesday, September 28, 1999.
1 Machine Learning: Lecture 6 Bayesian Learning (Based on Chapter 6 of Mitchell T.., Machine Learning, 1997)
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 24 of 41 Monday, 18 October.
Bayesian Learning Bayes Theorem MAP, ML hypotheses MAP learners
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Wednesday, 28 February 2007.
Computing & Information Sciences Kansas State University Monday, 09 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 19 of 42 Monday, 09 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 14 of 42 Wednesday, 22.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Computing & Information Sciences Kansas State University Wednesday, 04 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 17 of 42 Wednesday, 04 October.
Computing & Information Sciences Kansas State University Friday, 03 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 30 of 42 Friday, 03 November.
Bayesian Learning Evgueni Smirnov Overview Bayesian Theorem Maximum A Posteriori Hypothesis Naïve Bayes Classifier Learning Text Classifiers.
Computing & Information Sciences Kansas State University Friday, 13 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 21 of 42 Friday, 13 October.
Kansas State University Department of Computing and Information Sciences CIS 732: Machine Learning and Pattern Recognition Monday, 01 February 2016 William.
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Monday, 01 December 2003 William.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Monday, 28 November 2005 William.
CSC321: Lecture 8: The Bayesian way to fit models Geoffrey Hinton.
Computing & Information Sciences Kansas State University Wednesday, 25 Oct 2006CIS 490 / 730: Artificial Intelligence Lecture 26 of 42 Wednesday. 25 October.
Computing & Information Sciences Kansas State University Monday, 18 Sep 2006CIS 490 / 730: Artificial Intelligence Lecture 11 of 42 Monday, 18 September.
Bayesian Learning. Uncertainty & Probability Baye's rule Choosing Hypotheses- Maximum a posteriori Maximum Likelihood - Baye's concept learning Maximum.
Computing & Information Sciences Kansas State University Wednesday, 01 Nov 2006CIS 490 / 730: Artificial Intelligence Lecture 29 of 42 Wednesday, 01 November.
Naive Bayes Classifier
Data Mining Lecture 11.
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Naive Bayes Classifier
Presentation transcript:

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 25 Wednesday, 20 October 2004 William H. Hsu Department of Computing and Information Sciences, KSU Reading: Chapter 12, Russell and Norvig 2e Probability Background and Hour Exam 1 Review

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Making Decisions under Uncertainty Adapted from slides by S. Russell, UC Berkeley

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Bayes’s Theorem [1] Adapted from slides by S. Russell, UC Berkeley

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Bayes’s Theorem [2] Theorem P(h)  Prior Probability of Assertion (Hypothesis) h –Measures initial beliefs (BK) before any information is obtained (hence prior) P(D)  Prior Probability of Data (Observations) D –Measures probability of obtaining sample D (i.e., expresses D) P(h | D)  Probability of h Given D –| denotes conditioning - hence P(h | D) is a conditional (aka posterior) probability P(D | h)  Probability of D Given h –Measures probability of observing D given that h is correct (“generative” model) P(h  D)  Joint Probability of h and D –Measures probability of observing D and of h being correct

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Bayesian Inference: Query Answering (QA) Answering User Queries –Suppose we want to perform intelligent inferences over a database DB Scenario 1: DB contains records (instances), some “labeled” with answers Scenario 2: DB contains probabilities (annotations) over propositions –QA: an application of probabilistic inference QA Using Prior and Conditional Probabilities: Example –Query: Does patient have cancer or not? –Suppose: patient takes a lab test and result comes back positive Correct + result in only 98% of the cases in which disease is actually present Correct - result in only 97% of the cases in which disease is not present Only of the entire population has this cancer –   P(false negative for H 0  Cancer) = 0.02 (NB: for 1-point sample) –   P(false positive for H 0  Cancer) = 0.03 (NB: for 1-point sample) –P(+ | H 0 ) P(H 0 ) = , P(+ | H A ) P(H A ) =  h MAP = H A   Cancer

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Choosing Hypotheses Bayes’s Theorem MAP Hypothesis –Generally want most probable hypothesis given the training data –Define:  the value of x in the sample space  with the highest f(x) –Maximum a posteriori hypothesis, h MAP ML Hypothesis –Assume that p(h i ) = p(h j ) for all pairs i, j (uniform priors, i.e., P H ~ Uniform) –Can further simplify and choose the maximum likelihood hypothesis, h ML

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Terminology Introduction to Reasoning under Uncertainty –Probability foundations –Definitions: subjectivist, frequentist, logicist –(3) Kolmogorov axioms Bayes’s Theorem –Prior probability of an event –Joint probability of an event –Conditional (posterior) probability of an event Maximum A Posteriori (MAP) and Maximum Likelihood (ML) Hypotheses –MAP hypothesis: highest conditional probability given observations (data) –ML: highest likelihood of generating the observed data –ML estimation (MLE): estimating parameters to find ML hypothesis Bayesian Inference: Computing Conditional Probabilities (CPs) in A Model Bayesian Learning: Searching Model (Hypothesis) Space using CPs

Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Summary Points Introduction to Probabilistic Reasoning –Framework: using probabilistic criteria to search H –Probability foundations Definitions: subjectivist, objectivist; Bayesian, frequentist, logicist Kolmogorov axioms Bayes’s Theorem –Definition of conditional (posterior) probability –Product rule Maximum A Posteriori (MAP) and Maximum Likelihood (ML) Hypotheses –Bayes’s Rule and MAP –Uniform priors: allow use of MLE to generate MAP hypotheses –Relation to version spaces, candidate elimination Next Week: Chapter 15, Russell and Norvig –Later: Bayesian learning: MDL, BOC, Gibbs, Simple (Naïve) Bayes –Categorizing text and documents, other applications