An Algorithm to Learn the Structure of a Bayesian Network Çiğdem Gündüz Olcay Taner Yıldız Ethem Alpaydın Computer Engineering Taner Bilgiç Industrial.

Slides:



Advertisements
Similar presentations
CS188: Computational Models of Human Behavior
Advertisements

Bayesian network for gene regulatory network construction
A Tutorial on Learning with Bayesian Networks
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for 1 Lecture Notes for E Alpaydın 2010.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Variational Methods for Graphical Models Micheal I. Jordan Zoubin Ghahramani Tommi S. Jaakkola Lawrence K. Saul Presented by: Afsaneh Shirazi.
CS498-EA Reasoning in AI Lecture #15 Instructor: Eyal Amir Fall Semester 2011.
“Using Weighted MAX-SAT Engines to Solve MPE” -- by James D. Park Shuo (Olivia) Yang.
Lauritzen-Spiegelhalter Algorithm
Experiments We measured the times(s) and number of expanded nodes to previous heuristic using BFBnB. Dynamic Programming Intuition. All DAGs must have.
1 Knowledge Engineering for Bayesian Networks. 2 Probability theory for representing uncertainty l Assigns a numerical degree of belief between 0 and.
Introduction of Probabilistic Reasoning and Bayesian Networks
Learning Causality Some slides are from Judea Pearl’s class lecture
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Junction Trees: Motivation Standard algorithms (e.g., variable elimination) are inefficient if the undirected graph underlying the Bayes Net contains cycles.
Machine Learning CUNY Graduate Center Lecture 6: Junction Tree Algorithm.
GS 540 week 6. HMM basics Given a sequence, and state parameters: – Each possible path through the states has a certain probability of emitting the sequence.
Bayesian network inference
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Midterm Review. The Midterm Everything we have talked about so far Stuff from HW I won’t ask you to do as complicated calculations as the HW Don’t need.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Bayes Nets Rong Jin. Hidden Markov Model  Inferring from observations (o i ) to hidden variables (q i )  This is a general framework for representing.
Graphical Models Lei Tang. Review of Graphical Models Directed Graph (DAG, Bayesian Network, Belief Network) Typically used to represent causal relationship.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
PatReco: Bayesian Networks Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Today Logistic Regression Decision Trees Redux Graphical Models
Bayesian Networks Alan Ritter.
Causal Modeling for Anomaly Detection Andrew Arnold Machine Learning Department, Carnegie Mellon University Summer Project with Naoki Abe Predictive Modeling.
Machine Learning CUNY Graduate Center Lecture 21: Graphical Models.
A Brief Introduction to Graphical Models
Nirmalya Roy School of Electrical Engineering and Computer Science Washington State University Cpt S 223 – Advanced Data Structures Graph Algorithms: Minimum.
Undirected Models: Markov Networks David Page, Fall 2009 CS 731: Advanced Methods in Artificial Intelligence, with Biomedical Applications.
Spanning Trees Introduction to Spanning Trees AQR MRS. BANKS Original Source: Prof. Roger Crawfis from Ohio State University.
Spanning Trees Introduction to Spanning Trees AQR MRS. BANKS Original Source: Prof. Roger Crawfis from Ohio State University.
第十讲 概率图模型导论 Chapter 10 Introduction to Probabilistic Graphical Models
Daphne Koller Variable Elimination Graph-Based Perspective Probabilistic Graphical Models Inference.
Learning Linear Causal Models Oksana Kohutyuk ComS 673 Spring 2005 Department of Computer Science Iowa State University.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Announcements Project 4: Ghostbusters Homework 7
Computer Science: A Structured Programming Approach Using C Graphs A graph is a collection of nodes, called vertices, and a collection of segments,
Daphne Koller Message Passing Belief Propagation Algorithm Probabilistic Graphical Models Inference.
Slides for “Data Mining” by I. H. Witten and E. Frank.
An Introduction to Variational Methods for Graphical Models
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Lecture 2: Statistical learning primer for biologists
Learning and Acting with Bayes Nets Chapter 20.. Page 2 === A Network and a Training Data.
Christopher M. Bishop, Pattern Recognition and Machine Learning 1.
Neural Trees Olcay Taner Yıldız, Ethem Alpaydın Boğaziçi University Computer Engineering Department
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Introduction on Graphic Models
Today Graphical Models Representing conditional dependence graphically
1 Relational Factor Graphs Lin Liao Joint work with Dieter Fox.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016.
INTRODUCTION TO Machine Learning 2nd Edition
Inference in Bayesian Networks
Akbar Akbari Esfahani1, Theodor Asch2
Menglong Li Ph.d of Industrial Engineering Dec 1st 2016
Inconsistent Constraints
CAP 5636 – Advanced Artificial Intelligence
Markov Networks.
An Algorithm for Bayesian Network Construction from Data
Markov Random Fields Presented by: Vladan Radosavljevic.
CS 188: Artificial Intelligence
Markov Networks.
Presentation transcript:

An Algorithm to Learn the Structure of a Bayesian Network Çiğdem Gündüz Olcay Taner Yıldız Ethem Alpaydın Computer Engineering Taner Bilgiç Industrial Engineering Boğaziçi University

Bayesian Networks Graphical model to encode probabilistic relationships among data Consists of –Directed Acyclic Graph (DAG) –Conditional Probabilities

Example Bayesian Network

Issues in Bayesian Networks Given data, learning the structure of the Bayesian Network (NP-Complete) –Finding the arcs (dependencies) between the nodes –Calculating conditional probability tables Given the Bayesian Network, finding an efficient algorithm for inference on a given structure (NP-Complete)

Structure Learning Algorithms Based on a maximization of a measure –Likelihood Using Independence Criteria –Representing as much as the original dependencies in the data Hybrid of the former two –Our algorithm is in this group

Conditional Independence Variable X and Y are conditionally independent   X, Y and Z, P ( X | Y, Z ) = P ( X | Z ), whenever P ( Y, Z ) > 0 Cardinality of Z indicates the order of conditional independency

Our Algorithm Obtain the undirected graph using 0 and 1 independencies Find the ordering that minimizes the size of the conditional tables Using modify (change direction of the arc) and remove arc obtain final network Calculate conditional probability tables

Obtaining the Undirected Graph Find 0 and 1 independences using Mutual Information Test Add edges according to 0 and 1 independences until the graph is connected

Variable Ordering Algorithm For each variable –Assign all neighbor edges as incoming arcs –Compute size of the conditional tables –Mark variable as unselected While there are unselected nodes –Select the node with the minimum table size –Put the node in the ordering list –Mark node as selected –Adjust conditional table size of unselected nodes

Learning Steps Calculate likelihood of the data before and after applying the two operators on cv set If the operator improves the likelihood we accept that operator We continue until there is no improvement

Learning of a 4 node Network

Obtaining Undirected Graph

Obtaining DAG

Obtaining Conditional Tables

Results on the Original Alarm Network Original Graph has 46 arcs Our algorithm has only 3 missing arcs 11 arcs are inverted There are 23 extra arcs D-separation can also be used to remove unnecessary arcs

Alarm Network

Conclusion A novel algorithm for learning the structure of Bayesian Network is proposed Algorithm runs well in small networks –Similar likelihoods with the original network –Similar structures with the right directions The algorithm heavily depends on data as all 0 and 1 independence tests are based on a statistical test.

Future Work Missing variables can be filled with EM algorithm We can add further operators such as adding hidden nodes with appropriate arcs To check the validity of our algorithm we can use several classification data sets and use the model we learned to make classifications