BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
A Tutorial on Learning with Bayesian Networks
BAYESIAN NETWORKS Ivan Bratko Faculty of Computer and Information Sc. University of Ljubljana.
Probabilistic Reasoning Bayesian Belief Networks Constructing Bayesian Networks Representing Conditional Distributions Summary.
Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
Knowledge Representation and Reasoning University "Politehnica" of Bucharest Department of Computer Science Fall 2010 Adina Magda Florea
Artificial Intelligence Universitatea Politehnica Bucuresti Adina Magda Florea
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Identifying Conditional Independencies in Bayes Nets Lecture 4.
Bayesian Networks VISA Hyoungjune Yi. BN – Intro. Introduced by Pearl (1986 ) Resembles human reasoning Causal relationship Decision support system/ Expert.
1 22c:145 Artificial Intelligence Bayesian Networks Reading: Ch 14. Russell & Norvig.
Introduction of Probabilistic Reasoning and Bayesian Networks
Causal and Bayesian Network (Chapter 2) Book: Bayesian Networks and Decision Graphs Author: Finn V. Jensen, Thomas D. Nielsen CSE 655 Probabilistic Reasoning.
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graph.
Probabilistic Reasoning Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 14 (14.1, 14.2, 14.3, 14.4) Capturing uncertain knowledge Probabilistic.
PGM 2003/04 Tirgul 3-4 The Bayesian Network Representation.
Bayesian networks Chapter 14 Section 1 – 2.
Bayesian Belief Networks
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
A Differential Approach to Inference in Bayesian Networks - Adnan Darwiche Jiangbo Dang and Yimin Huang CSCE582 Bayesian Networks and Decision Graphs.
. DAGs, I-Maps, Factorization, d-Separation, Minimal I-Maps, Bayesian Networks Slides by Nir Friedman.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
Bayesian Networks Material used 1 Random variables
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition.
A Brief Introduction to Graphical Models
Bayesian networks Chapter 14 Section 1 – 2. Bayesian networks A simple, graphical notation for conditional independence assertions and hence for compact.
Bayesian networks. Motivation We saw that the full joint probability can be used to answer any question about the domain, but can become intractable as.
1 Chapter 14 Probabilistic Reasoning. 2 Outline Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions.
2 Syntax of Bayesian networks Semantics of Bayesian networks Efficient representation of conditional distributions Exact inference by enumeration Exact.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
1 Monte Carlo Artificial Intelligence: Bayesian Networks.
Introduction to Bayesian Networks
An Introduction to Artificial Intelligence Chapter 13 & : Uncertainty & Bayesian Networks Ramin Halavati
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
Bayesian Networks Aldi Kraja Division of Statistical Genomics.
1 Bayesian Networks (Directed Acyclic Graphical Models) The situation of a bell that rings whenever the outcome of two coins are equal can not be well.
Bayesian Networks CSE 473. © D. Weld and D. Fox 2 Bayes Nets In general, joint distribution P over set of variables (X 1 x... x X n ) requires exponential.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
Exploiting Structure in Probability Distributions Irit Gat-Viks Based on presentation and lecture notes of Nir Friedman, Hebrew University.
Lecture 29 Conditional Independence, Bayesian networks intro Ch 6.3, 6.3.1, 6.5, 6.5.1,
1 Use graphs and not pure logic Variables represented by nodes and dependencies by edges. Common in our language: “threads of thoughts”, “lines of reasoning”,
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
Introduction on Graphic Models
Conditional Probability, Bayes’ Theorem, and Belief Networks CISC 2315 Discrete Structures Spring2010 Professor William G. Tanner, Jr.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
Belief Networks Kostas Kontogiannis E&CE 457. Belief Networks A belief network is a graph in which the following holds: –A set of random variables makes.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
A Brief Introduction to Bayesian networks
Reasoning Under Uncertainty: Belief Networks
CS 2750: Machine Learning Directed Graphical Models
Bayesian networks Chapter 14 Section 1 – 2.
Presented By S.Yamuna AP/CSE
Learning Bayesian Network Models from Data
Bayesian Networks Based on
CS 188: Artificial Intelligence Fall 2007
Bayesian networks Chapter 14 Section 1 – 2.
Probabilistic Reasoning
Bayesian Networks CSE 573.
Probabilistic Reasoning
Chapter 14 February 26, 2004.
Presentation transcript:

BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009

Introduction  Joint Probability Distribution can be used to model uncertain beliefs and change them in the face of Hard and Soft Evidence.  Problem with JPD is that size grows exponentially with the number of variables which introduces modeling and computational difficulties.

Need for BN  BN is a graphical modeling tool for compactly specifying JPD  BN relies on the basic insight that: “ independence forms a significant aspect of belief” “Elicitation is relatively easily using the language of graph”

Example Earthquake (E) Burglary (B) Alarm (A) Radio (R) Call (C)  BN is a Directed Acyclic Graph Nodes are Proposition al Variables Edges are Direct Causal Influences

Example  We would expect our belief in C to be influenced by some Evidence on R  For example if we get a Radio report that an Earthquake took place then our belief in Alarm triggering would increase which would increase our belief in receiving call from a neighbor  However we would not change our belief if we knew for sure that the Alarm did not trigger  Thus C would be independent of R given ¬A

Formal Representation of Independence Given a variable V in a DAG G:  Parents (V) are the parents of V [Direct Causes of V]  Descendants(V) are the set of variables N with a directed path from V to N [Effects of V]  Non_Descendants(V) are the variables other that Parents and Descendants

Independence Statement / Markovian Assumption

Examples of Independence Statements  I (C,A, {B,E,R} )  I (R,E, {A,B,C} )  I (A,{B,E}, R)  I (B, ø, {E,R})  I (E, ø, B) Earthquake (E) Burglary (B) Alarm (A) Radio (R) Call (C)

Parameterizing the Independence Structure  Parameterizing means quantifying the dependencies between Nodes and their Parents  In other words construction of CPT  For every variable X in the DAG G and its parents U, we need to provide the probability Pr(x|u) for every value x of variable X and every instantiation u of parents U

Formal Definition of Bayesian Network  A Bayesian Network for variables Z is a pair where:  G is a directed acyclic graph over variables Z called the Network Structure  is a set of CPT’s one for each variable in Z called the Network Parameterization  (X|U) would be used to denote the CPT for variable X and its parents U, and refer to the set XU as a Network Family.

Def (continue..)  denotes the value assigned by CPT to the conditional probability Pr (x|u) and call it Network Parameter  Instantiation of all the network variables are called Network Instantiations Network parameterNetwork instantiation a a a  (b|a) b  ( ¬ c|a) ¬c  (d|b, ¬ c) d  ( ¬ e| ¬ c) ¬e

Chain Rule for Bayesian Networks  Network Instantiations z is simply the product of all network parameters compatible with z

Properties of Probabilistic Independence  Recall : I (X,Z,Y) Pr(x|z,y) = Pr(x|z) or Pr(y|z) =0 for all instantiations x,y,z  Graphoid Axioms:  Symmetry  Weak Union  Decomposition  Contraction

Symmetry  If learning Y does not influence our belief in x then learning x does not influence our belief in y  By Markov(G) we know that: I (A,{B,E},R) Using Symmetry: I (R,{B,E},A) Earthquake (E) Burglary (B) Alarm (A) Radio (R) Call (C)

Decomposition  If learning yw does not influence our belief in x then learning y alone or learning w alone does not influence our belief in x

Weak Union  If the information yw is not relevant to our belief in x then the partial information will not make the rest of the information relevant

Contraction  If learning the irrelevant information y the information w is found to be irrelevant to our belief in x then the combined information must have been irrelevant from the beginning

Questions ???