COMP 538 Reasoning and Decision under Uncertainty Introduction Readings: Pearl (1998, Chapter 1 Shafer and Pearl, Chapter 1.

Slides:



Advertisements
Similar presentations
1 Probability and the Web Ken Baclawski Northeastern University VIStology, Inc.
Advertisements

CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 27 – Overview of probability concepts 1.
Graphical Models and Applications CNS/EE148 Instructors: M.Polito, P.Perona, R.McEliece TA: C. Fanti.
Learning with Missing Data
Graphical Models - Inference - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
Learning: Parameter Estimation
Introduction to Belief Propagation and its Generalizations. Max Welling Donald Bren School of Information and Computer and Science University of California.
Graphical Models - Learning -
Bayesian Networks - Intro - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP HREKG.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
Graphical Models - Inference -
Graphical Models - Learning - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
CS B553: A LGORITHMS FOR O PTIMIZATION AND L EARNING aka “Neural and Genetic Approaches to Artificial Intelligence” Spring 2011 Kris Hauser.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
Graphical Models - Inference - Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany PCWP CO HRBP.
. PGM: Tirgul 10 Learning Structure I. Benefits of Learning Structure u Efficient learning -- more accurate models with less data l Compare: P(A) and.
COMP 538 (Reasoning and Decision under Uncertainty) Introduction to Bayesian networks Introduction to Course.
CPSC 322, Lecture 31Slide 1 Probability and Time: Markov Models Computer Science cpsc322, Lecture 31 (Textbook Chpt 6.5) March, 25, 2009.
CPSC 322, Lecture 32Slide 1 Probability and Time: Hidden Markov Models (HMMs) Computer Science cpsc322, Lecture 32 (Textbook Chpt 6.5) March, 27, 2009.
CPSC 422, Lecture 14Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14 Feb, 4, 2015 Slide credit: some slides adapted from Stuart.
CSE 590ST Statistical Methods in Computer Science Instructor: Pedro Domingos.
CIS 410/510 Probabilistic Methods for Artificial Intelligence Instructor: Daniel Lowd.
CSE 515 Statistical Methods in Computer Science Instructor: Pedro Domingos.
Medical Informatics Basics
CS Machine Learning. What is Machine Learning? Adapt to / learn from data  To optimize a performance function Can be used to:  Extract knowledge.
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
A Brief Introduction to Graphical Models
Last Words COSC Big Data (frameworks and environments to analyze big datasets) has become a hot topic; it is a mixture of data analysis, data mining,
Anomaly detection with Bayesian networks Website: John Sandiford.
Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Overview Particle filtering is a sequential Monte Carlo methodology in which the relevant probability distributions are iteratively estimated using the.
Introduction to Artificial Intelligence and Soft Computing
Model-based Bayesian Reinforcement Learning in Partially Observable Domains by Pascal Poupart and Nikos Vlassis (2008 International Symposium on Artificial.
Last Words DM 1. Mining Data Steams / Incremental Data Mining / Mining sensor data (e.g. modify a decision tree assuming that new examples arrive continuously,
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Real-Time AI Wednesday, March 14, 2001 Haipeng Guo.
Notes on Graphical Models Padhraic Smyth Department of Computer Science University of California, Irvine.
AI ● Dr. Ahmad aljaafreh. What is AI? “AI” can be defined as the simulation of human intelligence on a machine, so as to make the machine efficient to.
CS Statistical Machine learning Lecture 24
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CS498-EA Reasoning in AI Lecture #10 Instructor: Eyal Amir Fall Semester 2009 Some slides in this set were adopted from Eran Segal.
Artificial Intelligence, Expert Systems, and Neural Networks Group 10 Cameron Kinard Leaundre Zeno Heath Carley Megan Wiedmaier.
CH751 퍼지시스템 특강 Uncertainties in Intelligent Systems 2004 년도 제 1 학기.
Tractable Inference for Complex Stochastic Processes X. Boyen & D. Koller Presented by Shiau Hong Lim Partially based on slides by Boyen & Koller at UAI.
Some Neat Results From Assignment 1. Assignment 1: Negative Examples (Rohit)
Motivation and Overview
Guidance: Assignment 3 Part 1 matlab functions in statistics toolbox  betacdf, betapdf, betarnd, betastat, betafit.
1 CSC 384 Lecture Slides (c) , C. Boutilier and P. Poupart CSC384: Intro to Artificial Intelligence  Fahiem Bacchus Office.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
CS382 Introduction to Artificial Intelligence Lecture 1: The Foundations of AI and Intelligent Agents 24 January 2012 Instructor: Kostas Bekris Computer.
From NARS to a Thinking Machine Pei Wang Temple University.
Daphne Koller Introduction Motivation and Overview Probabilistic Graphical Models.
Learning Bayesian Networks for Complex Relational Data
CHAPTER 1 Introduction BIC 3337 EXPERT SYSTEM.
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Introduction to Artificial Intelligence
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 14
Course: Autonomous Machine Learning
Introduction to Artificial Intelligence
Data Mining: Concepts and Techniques Course Outline
Read R&N Ch Next lecture: Read R&N
Uncertainty in AI.
Introduction to Artificial Intelligence and Soft Computing
Introduction to Artificial Intelligence
CSE 515 Statistical Methods in Computer Science
Probability and Time: Markov Models
Probability and Time: Markov Models
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Presentation transcript:

COMP 538 Reasoning and Decision under Uncertainty Introduction Readings: Pearl (1998, Chapter 1 Shafer and Pearl, Chapter 1

COMP 538 Introduction / Slide 2 Objectives n Course objectives n Course contents

COMP 538 Introduction / Slide 3 Uncertainty  Uncertainty: the quality or state of being not clearly known.  Uncertainty appears in many tasks  Partial knowledge of the state of the world  Noisy observations  Phenomena that is not covered by our models  Inherent randomness

COMP 538 Introduction / Slide 4 Probability and Decision Theory  Well-known and well-understood framework for uncertainty  Clear semantics  Provides principled answers for:  Combining evidence  Predictive & Diagnostic reasoning  Incorporation of new evidence  Intuitive (at some level) to human experts  Can be learned

COMP 538 Introduction / Slide 5 Course Objectives n When applied to real-world problems, probability theory and decision theory suffer from n Complexity of model construction n Complexity of problem solving n This course covers methodologies developed recently in AI community for dealing with those complexity problems. n The methodologies combine ideas from several disciplines Artificial Intelligence, Machine Learning Decision Theory, Theory of Computer Science Statistics, Information Theory, Operations Research

COMP 538 Introduction / Slide 6 Complexity Problem of Applying Probability Theory n Example: n Patients in hospital are described by several attributes: n Background: age, gender, history of diseases, … n Symptoms: fever, blood pressure, headache, … n Diseases: pneumonia, heart attack, … n A joint probability distribution needs to assign a number to each combination of values of these attributes, exponential model size. 20 attributes require ( roughly 10 6 ) numbers n Real applications usually involve hundreds of attributes

COMP 538 Introduction / Slide 7 n Because of the exponential model size problem, it was believed that probability theory is not practical for dealing with uncertainty in AI. n Alternative uncertainty calculi were introduced: uncertainty factors, non-monotonic logic, fuzzy logic, etc. Complexity Problem of Applying Probability Theory

COMP 538 Introduction / Slide 8 n Bayesian networks alleviate the exponential model size problem. n Key idea: use conditional independence to factorize model into smaller parts. n Example: Alarm network 37 variables Model size: 2 37 Size of factored model: 509 PCWP CO HRBP HREKG HRSAT ERRCAUTER HR HISTORY CATECHOL SAO2 EXPCO2 ARTCO2 VENTALV VENTLUNG VENITUBE DISCONNECT MINVOLSET VENTMACH KINKEDTUBE INTUBATIONPULMEMBOLUS PAPSHUNT ANAPHYLAXIS MINOVL PVSAT FIO2 PRESS INSUFFANESTHTPR LVFAILURE ERRBLOWOUTPUT STROEVOLUMELVEDVOLUME HYPOVOLEMIA CVP BP Complexity Problem of Applying Probability Theory Model construction and problem solving becomes possible

COMP 538 Introduction / Slide 9 Advantages of Bayesian Networks n Semantics n Probability theory provides the glue whereby the parts are combined, ensuring that the system as a whole is consistent. n Alternative approaches suffer from several semantic deficiencies (Pearl 1988, Chapter 1). n Model construction by expert n Appealing graphical interface allows experts to build model for highly interacting variables. n Model construction from data n Probability foundation allows model construction from data by well established statistical principles such as maximum likelihood estimation and Bayesian estimation.

COMP 538 Introduction / Slide 10 Fielded Applications  Expert systems  Medical diagnosis  Fault diagnosis (jet-engines, Windows 98)  Monitoring  Space shuttle engines (Vista project)  Freeway traffic  Sequence analysis and classification  Speech recognition  Biological sequences  Information access  Collaborative filtering  Information retrieval See tutorial by Breese and Koller (1997) and online resources for application samples.

COMP 538 Introduction / Slide 11 Course Content/Bayesian Networks n Concept and semantics of Bayesian networks n Inference: How to answer queries efficiently n Learning: How to learn/adapt Bayesian network models from data n Causal models: How to learn causality from statistical data Detailed discussion of those topics can take one whole course (DUKE). We will focus on the main ideas and skip the details so that we can study other related topics.

COMP 538 Introduction / Slide 12 n Special cases of Bayesian networks: many of the classical multivariate models from n statistics, systems engineering, information theory, pattern recognition and statistical mechanics n Examples: mixture models, factor analysis, hidden Markov models, Kalman filters, Ising models. n Bayesian networks provide a way to view all of these models as instances of a common underlying formalism. Bayesian networks and classical multivariate models

COMP 538 Introduction / Slide 13 Course Content/Special models n Latent class analysis: n Statistical method for finding subtypes of related cases. n With proper generalization, might provide a statistical foundation for Chinese medicine diagnosis. n Hidden Markov models: n A temporal model widely used in pattern recognition: handwriting recognition, speech recognition.

COMP 538 Introduction / Slide 14 Decision Making under Uncertainty n Typical scenario: whether to take umbrella n Decision theory provides basis for rational decision making: Maximum expected utility principle. n Decision analysis: applying of decision theory. n Suffers from exponential model size.

COMP 538 Introduction / Slide 15 Course Content/Simple Decision Making n Influence diagrams n Generalization of Bayesian network. n Alleviate the complexity problem of decision analysis n Topics: n Evaluation: How to compute optimal decisions in an influence diagram. n Value of information: Whether it is worthwhile to collect more information to reduce uncertainty.

COMP 538 Introduction / Slide 16 n Agent/robot needs to execute multiple actions in order to achieve a goal n Uncertainty originates from noisy sensor and inaccurate actuators/uncontrollable enviroment factors n What is the best way to reach goal with minimum cost/time? Sequential Decision Making under Uncertainty

COMP 538 Introduction / Slide 17 Course Content/Sequential Decision Making n Models: n Markov decision processes, consider only uncertainty in actuators/environment factors n Partially observable decision processes, consider uncertainty in sensors and actuators/environment factors. n Solution methods: n Value iteration n Policy iteration n Dealing with model complexity using dynamic Bayesian networks n Learning sequential decision models n Model-based reinforcement learning

COMP 538 Introduction / Slide 18 Course Content/Summary n Bayesian networks n Concept and semantics, inference, learning, causilty n Special models: Hidden Markov models, latent class analysis n Influence diagrams: n Evaluation, value of information n Markov decision processes and partially observable Markov decision processes n Solution methods: value iteration, policy iteration, dynamic Bayesian networks n Model-based reinforcement learning