Introduction to Bayesian Networks

Slides:



Advertisements
Similar presentations
Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Advertisements

BAYESIAN NETWORKS Ivan Bratko Faculty of Computer and Information Sc. University of Ljubljana.
Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Slide 1 Reasoning Under Uncertainty: More on BNets structure and construction Jim Little Nov (Textbook 6.3)
For Monday Finish chapter 14 Homework: –Chapter 13, exercises 8, 15.
Bayesian Networks VISA Hyoungjune Yi. BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert.
Bayesian Networks Chapter 14 Section 1, 2, 4. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Introduction of Probabilistic Reasoning and Bayesian Networks
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) March, 16, 2009.
M.I. Jaime Alfonso Reyes ´Cortés.  The basic task for any probabilistic inference system is to compute the posterior probability distribution for a set.
Review: Bayesian learning and inference
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
Bayesian Networks. Motivation The conditional independence assumption made by naïve Bayes classifiers may seem to rigid, especially for classification.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
PGM 2003/04 Tirgul 3-4 The Bayesian Network Representation.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
. DAGs, I-Maps, Factorization, d-Separation, Minimal I-Maps, Bayesian Networks Slides by Nir Friedman.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
1 CMSC 471 Fall 2002 Class #19 – Monday, November 4.
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Bayesian Networks What is the likelihood of X given evidence E? i.e. P(X|E) = ?
Making Simple Decisions
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
For Wednesday Read Chapter 11, sections 1-2 Program 2 due.
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
CS 188: Artificial Intelligence Fall 2006 Lecture 18: Decision Diagrams 10/31/2006 Dan Klein – UC Berkeley.
Probabilistic Networks Chapter 14 of Dechter’s CP textbook Speaker: Daniel Geschwender April 1, 2013 April 1&3, 2013DanielG--Probabilistic Networks1.
Bayesian Networks CSE 473. © D. Weld and D. Fox 2 Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
Exploiting Structure in Probability Distributions Irit Gat-Viks Based on presentation and lecture notes of Nir Friedman, Hebrew University.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
1 Probability FOL fails for a domain due to: –Laziness: too much to list the complete set of rules, too hard to use the enormous rules that result –Theoretical.
Introduction on Graphic Models
Belief Networks Kostas Kontogiannis E&CE 457. Belief Networks A belief network is a graph in which the following holds: –A set of random variables makes.
CPSC 322, Lecture 26Slide 1 Reasoning Under Uncertainty: Belief Networks Computer Science cpsc322, Lecture 27 (Textbook Chpt 6.3) Nov, 13, 2013.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
CS 2750: Machine Learning Review
CS 2750: Machine Learning Directed Graphical Models
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Qian Liu CSE spring University of Pennsylvania
Read R&N Ch Next lecture: Read R&N
Learning Bayesian Network Models from Data
CSCI 121 Special Topics: Bayesian Networks Lecture #2: Bayes Nets
Read R&N Ch Next lecture: Read R&N
CAP 5636 – Advanced Artificial Intelligence
Bayesian Statistics and Belief Networks
Class #21 – Monday, November 10
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
Class #16 – Tuesday, October 26
Belief Networks CS121 – Winter 2003 Belief Networks.
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Read R&N Ch Next lecture: Read R&N
Bayesian Networks CSE 573.
Probabilistic Reasoning
Chapter 14 February 26, 2004.
Presentation transcript:

Introduction to Bayesian Networks

Joint Probability Distribution Representation: What is the joint probability distribution P(B, E, R, A, N) over five binary variables? How many states in total? How many free parameters do we need? How many table entries should we sum over to get P(R=r)?

Joint Probability Distribution Full joint distribution is sufficient to represent the complete domain and to do any type of probabilistic inferences Problems: n - number of random variables, d – number of values Space complexity. A full joint distribution requires to remember O(dn) numbers Inference complexity. Computing queries requires O(dn) steps Acquisition problem. Who is going to define all the probabilities?

Joint Probability Distribution

Joint Probability Distribution

Joint Probability Distribution VS

Bayesian networks A Bayesian Network is a graph in which: P(E) P(B) A set of random variables makes up the nodes in the network. A set of directed links or arrows connects pairs of nodes. Each node has a conditional probability table that quantifies the effects the parents have on the node. Directed, acyclic graph (DAG), i.e. no directed cycles. P(E) P(B) A Bayesian network modeling the joint probability distribution P(B, E, R, A, N) P(A|B,E) P(R|E) P(N|A)

Bayesian Networks Conditional probability tables (CPT): P(B) P(E) P(A|B,E) P(R|E) P(N|A)

Independences in BNs Three basic independence structures:

Independences in BNs Indirect cause: Burglary is independent of NeighborCall given Alarm P(N|A, B) = P(N|A) P(N, B|A) = P(N|A)P(B|A)

Independences in BNs Common cause: Alarm is independent of RadioAnnounce given Earthquake P(A|E, R) = P(A|E) P(A, R|E) = P(A|E)P(R|E)

Independences in BNs Common effect: Burglary is independent of Earthquake when Alarm is not known Burglary and Earthquake become dependent given Alarm!!!

Path blocking With linear substructure With common cause structure With common effect structure X Y C in Z X Y C in Z X Y C or any of its descendants not in Z

Independences in BNs Earthquake ┴ Burglary | NeighborCall? Burglary ┴ RadioAnnounce | Earthquake? Burglary ┴ RadioAnnounce | NeighborCall?

Markov Assumption Each variable is independent on its non-descendants, given its parents in Bayesian networks N ┴ B, E, R | A R ┴ B, A, N | E …

Full Joint Distribution in BNs ) , ( N R A E B P ) ( , | B P E A R N =

Chain Rule

Types of Inference Tasks Compute the probability or most likely state of a set of query variables given observed values of some other variables Belief updating Most probable explanation (MPE) queries Maximum A Posteriori (MAP) queries

Inference Direction Types of inference by reasoning direction

Examples Diagnostic inferences: from effect to causes. Causal Inferences: from causes to effects. Intercausal Inferences: Mixed Inference:

Influence diagram Add action nodes and utility nodes to Bayesian networks to enable rational decision making Algorithm: For each value of action node compute expected value of utility node given action, evidence Return MEU action

Value of information Idea: compute value of acquiring each possible piece of evidence Can be done directly from influence diagram Example: buying oil drilling rights Two blocks A and B, exactly one has oil, worth k Prior probabilities 0.5 each, mutually exclusive Current price of each block is k/2 “Consultant” offers accurate survey of A. Fair price? Solution: compute expected value of information = expected value of best action given the information minus expected value of best action without information Survey may say “oil in A” or “no oil in A”, prob. 0.5 each (given!) = [0.5 * value of “buy A” given “oil in A” + 0.5 * value of “buy B” given “no oil in A”] - 0 = (0.5 * k/2) + (0.5 * k/2) - 0 = k/2 Survey Oil Buy U