Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Bayesian networks Chapter 14 Section 1 – 2. Outline Syntax Semantics Exact computation.
Markov Networks Alan Ritter.
A Tutorial on Learning with Bayesian Networks
Bayesian Networks CSE 473. © Daniel S. Weld 2 Last Time Basic notions Atomic events Probabilities Joint distribution Inference by enumeration Independence.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
BAYESIAN NETWORKS. Bayesian Network Motivation  We want a representation and reasoning system that is based on conditional independence  Compact yet.
Supervised Learning Recap
BAYESIAN NETWORKS CHAPTER#4 Book: Modeling and Reasoning with Bayesian Networks Author : Adnan Darwiche Publisher: CambridgeUniversity Press 2009.
Introduction of Probabilistic Reasoning and Bayesian Networks
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Review: Bayesian learning and inference
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Bayesian Network Representation Continued
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
1 gR2002 Peter Spirtes Carnegie Mellon University.
Today Logistic Regression Decision Trees Redux Graphical Models
© Daniel S. Weld 1 Statistical Learning CSE 573 Lecture 16 slides which overlap fix several errors.
Bayes Net Perspectives on Causation and Causal Inference
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
Bayes’ Nets  A Bayes’ net is an efficient encoding of a probabilistic model of a domain  Questions we can ask:  Inference: given a fixed BN, what is.
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
A Brief Introduction to Graphical Models
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Soft Computing Lecture 17 Introduction to probabilistic reasoning. Bayesian nets. Markov models.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models
1 Robot Environment Interaction Environment perception provides information about the environment’s state, and it tends to increase the robot’s knowledge.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Tetrad project 
Introduction to Bayesian Networks
Methodological Problems in Cognitive Psychology David Danks Institute for Human & Machine Cognition January 10, 2003.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Course files
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Announcements Project 4: Ghostbusters Homework 7
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS.
Bayesian Network By Zhang Liliang. Key Point Today Intro to Bayesian Network Usage of Bayesian Network Reasoning BN: D-separation.
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Lecture notes 9 Bayesian Belief Networks.
INTERVENTIONS AND INFERENCE / REASONING. Causal models  Recall from yesterday:  Represent relevance using graphs  Causal relevance ⇒ DAGs  Quantitative.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Lecture 2: Statistical learning primer for biologists
The Visual Causality Analyst: An Interactive Interface for Causal Reasoning Jun Wang, Stony Brook University Klaus Mueller, Stony Brook University, SUNY.
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
1 BN Semantics 2 – Representation Theorem The revenge of d-separation Graphical Models – Carlos Guestrin Carnegie Mellon University September 17.
A Cooperative Coevolutionary Genetic Algorithm for Learning Bayesian Network Structures Arthur Carvalho
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
1 BN Semantics 1 Graphical Models – Carlos Guestrin Carnegie Mellon University September 15 th, 2006 Readings: K&F: 3.1, 3.2, 3.3.
Fast Exact Bayes Net Structure Learning Daniel Eaton Tuesday Oct 31, 2006 relatively-speakingly-
Graphical Models for Psychological Categorization David Danks Carnegie Mellon University; and Institute for Human & Machine Cognition.
Variable selection in Regression modelling Simon Thornley.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
Artificial Intelligence Bayes’ Nets: Independence Instructors: David Suter and Qince Li Course Harbin Institute of Technology [Many slides.
A Brief Introduction to Bayesian networks
CS 2750: Machine Learning Review
CS 2750: Machine Learning Directed Graphical Models
CS 4/527: Artificial Intelligence
A Bayesian Approach to Learning Causal networks
CAP 5636 – Advanced Artificial Intelligence
General Gibbs Distribution
An Algorithm for Bayesian Network Construction from Data
Class #19 – Tuesday, November 3
CS 188: Artificial Intelligence
Graduate School of Information Sciences, Tohoku University
CS 188: Artificial Intelligence Spring 2007
Chapter 14 February 26, 2004.
Presentation transcript:

Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition

Bayesian Networks Two components: 1. Directed Acyclic Graph (DAG) G: There is a node for every variable D: Some nodes have arrows btw. them (X  Y) A: Can’t get from one node back to itself by following arrows 2. Joint Probability Distribution For all {X,Y,…,Z}, P(X=x,Y=y,…,Z=z) is defined

Example of a Bayes Net Directed Acyclic Graph: Joint Probability Distribution: P(AP=High, B=High, ST=Yes) = 0.2 P(AP=High, B=Low, ST=Yes) = …

Connecting the Graph & JPD Markov assumption: “X is (probabilistically) independent of its (graphical) non-descendants conditional on its (graphical) parents.” “No edge  Conditional independence”

Connecting the Graph & JPD Markov assumption: “X is (probabilistically) independent of its (graphical) non-descendants conditional on its (graphical) parents.” “No edge  Conditional independence” Example: X  Y  Z  X Z | Y

Connecting the Graph & JPD The Markov assumption implies a factorization of the JPD based on the graph: P(X=x,Y=y,…,Z=z) =  P(v | parents(v)) The Markov assumption allows us to move from graph to probability distribution

Connecting the Graph & JPD Faithfulness assumption: “The (probabilistic) effects of (graphical) paths never exactly offset.” “Conditional independence  No edge” The Faithfulness assumption allows us to move from probability distribution to graph

Bayesian Network Example 1. Running causes you to eat more 2. Eating more causes you to gain weight 3. Running increases your metabolism 4. Increased metabolism leads to weight loss Note: Faithfulness rules out “Running Weight”

Learning Bayes Nets Given some data from the world, why would we want to learn a Bayes net? 1. Compact representation of the data There are fast algorithms for prediction/inference given observations of the environment 2. Causal knowledge There are fast algorithms for prediction/inference given interventions in the environment

Bayesian Updating Start with a probability distribution over possible graphs

Bayesian Updating Start with a probability distribution over possible graphs, then figure out which graph is most likely given the observed data

Features of Bayesian Updating Advantages: 1. Output is fine-grained probabilistic information 2. “Rational” basis for the learning algorithm 3. Robust to data errors Disadvantages: 1. Number of possible graphs is super-exponential, & likelihood functions not always solvable analytically  almost always use heuristics 2. Cannot easily incorporate unobserved variables

Constraint-Based Learning 1. Determine the (un)conditional associations and independencies in the data 2. Determine the set of Bayes nets that could have produced that data

Constraint-Based Learning 1. Determine the (un)conditional associations and independencies in the data 2. Determine the set of Bayes nets that could have produced that data X W (all) Y Z (all) X Y,Z | W Z X,W | Y

Constraint-Based Learning 1. Determine the (un)conditional associations and independencies in the data 2. Determine the set of Bayes nets that could have produced that data X W (all) Y Z (all) X Y,Z | W Z X,W | Y X WY Z U

Features of C-B Learning Advantages: 1. Feasible, asymptotically correct algorithms (though worst case requires exponential comp.) 2. Easily incorporates unobserved variables 3. Gives exactly the information in the data Disadvantages: 1. Susceptible to mistaken independence judgments (big problem for small datasets) 2. Cannot use fine-grained prior knowledge

Layers of “Bayes Nets” Start with simple association graph

Layers of “Bayes Nets” Then we can add a causal interpretation:

Layers of “Bayes Nets” Or we can provide parameter information: …

Layers of “Bayes Nets” Or we can combine causation & parameters …

Useful Websites Tutorials: Constraint-based learning software: (free) Bayesian learning software: (commercial) (free)