MCMC in structure space MCMC in order space.

Slides:



Advertisements
Similar presentations
When Efficient Model Averaging Out-Perform Bagging and Boosting Ian Davidson, SUNY Albany Wei Fan, IBM T.J.Watson.
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Dynamic Causal Modelling for ERP/ERFs Valentina Doria Georg Kaegi Methods for Dummies 19/03/2008.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
The IMAP Hybrid Method for Learning Gaussian Bayes Nets Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada
Markov-Chain Monte Carlo
PAC Learning adapted from Tom M.Mitchell Carnegie Mellon University.
Beam Sampling for the Infinite Hidden Markov Model Van Gael, et al. ICML 2008 Presented by Daniel Johnson.
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Estimation of Distribution Algorithms Ata Kaban School of Computer Science The University of Birmingham.
J. Daunizeau Wellcome Trust Centre for Neuroimaging, London, UK Institute of Empirical Research in Economics, Zurich, Switzerland Bayesian inference.
Particle filters (continued…). Recall Particle filters –Track state sequence x i given the measurements ( y 0, y 1, …., y i ) –Non-linear dynamics –Non-linear.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
Monte Carlo methods for estimating population genetic parameters Rasmus Nielsen University of Copenhagen.
CS 188: Artificial Intelligence Spring 2007 Lecture 14: Bayes Nets III 3/1/2007 Srini Narayanan – ICSI and UC Berkeley.
CS 188: Artificial Intelligence Fall 2006 Lecture 17: Bayes Nets III 10/26/2006 Dan Klein – UC Berkeley.
Finding the optimal pairwise alignment We are interested in finding the alignment of two sequences that maximizes the similarity score given an arbitrary.
Learning Bayesian Networks
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Bayesian Analysis for Extreme Events Pao-Shin Chu and Xin Zhao Department of Meteorology School of Ocean & Earth Science & Technology University of Hawaii-
Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks Dirk Husmeier Adriano V. Werhli.
Active Learning for Networked Data Based on Non-progressive Diffusion Model Zhilin Yang, Jie Tang, Bin Xu, Chunxiao Xing Dept. of Computer Science and.
Bayes Factor Based on Han and Carlin (2001, JASA).
Particle Filtering in Network Tomography
Mean Field Inference in Dependency Networks: An Empirical Study Daniel Lowd and Arash Shamaei University of Oregon.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Reverse Engineering of Genetic Networks (Final presentation)
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Charu Aggarwal + * Department of Computer Science, University of Texas at Dallas + IBM T. J. Watson.
Bayesian Learning By Porchelvi Vijayakumar. Cognitive Science Current Problem: How do children learn and how do they get it right?
Ahsanul Haque *, Swarup Chandra *, Latifur Khan * and Michael Baron + * Department of Computer Science, University of Texas at Dallas + Department of Mathematical.
Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Reverse engineering gene regulatory networks Dirk Husmeier Adriano Werhli Marco Grzegorczyk.
Learning regulatory networks from postgenomic data and prior knowledge Dirk Husmeier 1) Biomathematics & Statistics Scotland 2) Centre for Systems Biology.
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
Xiaowei Ying, Xintao Wu Univ. of North Carolina at Charlotte PAKDD-09 April 28, Bangkok, Thailand On Link Privacy in Randomizing Social Networks.
Inference Complexity As Learning Bias Daniel Lowd Dept. of Computer and Information Science University of Oregon Joint work with Pedro Domingos.
Collaborative Sampling in Wireless Sensor Networks Minglei Huang Yu Hen Hu 2010 IEEE Global Telecommunications Conference 1.
Simulation techniques Summary of the methods we used so far Other methods –Rejection sampling –Importance sampling Very good slides from Dr. Joo-Ho Choi.
Understanding Sampling
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
Problem Limited number of experimental replications. Postgenomic data intrinsically noisy. Poor network reconstruction.
Stable Multi-Target Tracking in Real-Time Surveillance Video
Inferring gene regulatory networks with non-stationary dynamic Bayesian networks Dirk Husmeier Frank Dondelinger Sophie Lebre Biomathematics & Statistics.
CHAPTER 5 Probability Theory (continued) Introduction to Bayesian Networks.
Learning Bayesian networks from postgenomic data with an improved structure MCMC sampling scheme Dirk Husmeier Marco Grzegorczyk 1) Biomathematics & Statistics.
Selecting Genomes for Reconstruction of Ancestral Genomes Louxin Zhang Department of Mathematics National University of Singapore.
Reverse engineering of regulatory networks Dirk Husmeier & Adriano Werhli.
Bayesian Optimization Algorithm, Decision Graphs, and Occam’s Razor Martin Pelikan, David E. Goldberg, and Kumara Sastry IlliGAL Report No May.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Introduction: Metropolis-Hasting Sampler Purpose--To draw samples from a probability distribution There are three steps 1Propose a move from x to y 2Accept.
A Cooperative Coevolutionary Genetic Algorithm for Learning Bayesian Network Structures Arthur Carvalho
Computational methods for inferring cellular networks II Stat 877 Apr 17 th, 2014 Sushmita Roy.
Hierarchical Mixture of Experts Presented by Qi An Machine learning reading group Duke University 07/15/2005.
CSC Lecture 23: Sigmoid Belief Nets and the wake-sleep algorithm Geoffrey Hinton.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Incorporating graph priors in Bayesian networks
CS 4/527: Artificial Intelligence
Markov Networks.
Multidimensional Integration Part I
true graph initial network keep edge and so on move to next edge 17.
CS 188: Artificial Intelligence
CS 188: Artificial Intelligence Fall 2008
CS639: Data Management for Data Science
Classical regression review
Presentation transcript:

MCMC in structure space

MCMC in order space

Current work with Marco Grzegorczyk MCMC in structure rather than order space. Design new proposal moves that achieve faster mixing and convergence.

First idea Propose new parents from the distribution: Identify those new parents that are involved in the formation of directed cycles. Orphan them, and sample new parents for them subject to the acyclicity constraint.

Problem: This move is not reversible

Design a complementary backward move, which proposes “illegal” strucures. Select a node X. Select a subset of its parents, propose new parents for these parents such that you get directed cycles that involve node X. Orphan node X, then select new parents subject to the acylicity constraint.

Move reversible, but maths complicated

Devise a simpler move with similar mixing and convergence Identify a pair of nodes X Y Orphan both nodes. Sample new parents from the Boltzmann distribution subject to the acyclicity constraint such the inverse edge Y X is included.

This move is reversible!

Acceptance probability

Does the new method avoid the bias intrinsic to order MCMC? How do convergence and mixing compare to structure and order MCMC? What is the effect on the network reconstruction accuracy?

Does the new method avoid the bias intrinsic to order MCMC? How do convergence and mixing compare to structure and order MCMC? What is the effect on the network reconstruction accuracy?

Estimating the bias of the method Consider a small network with only five nodes. A complete enumeration of structures is possible to compute the correct posterior distribution. Compute the difference between the predicted and the true marginal posterior probability, for all edges

Does the new method avoid the bias intrinsic to order MCMC? How do convergence and mixing compare to structure and order MCMC? What is the effect on the network reconstruction accuracy?

Alarm network Devised by Beinlich et al., 1989 N=37 nodes 46 directed edges We generated data sets with m=25,50,100,250,500,750,1000 instances

Does the new method avoid the bias intrinsic to order MCMC? How do convergence and mixing compare to structure and order MCMC? What is the effect on the network reconstruction accuracy?

Does the new method avoid the bias intrinsic to order MCMC? How do convergence and mixing compare to structure and order MCMC? What is the effect on the network reconstruction accuracy?

The new method avoids the bias intrinsic to order MCMC. Its convergence and mixing are similar to order MCMC; both methods outperform structure MCMC. Its network reconstruction accuracy is similar to order MCMC; both methods outperform structure MCMC. We expect to get an improvement over order MCMC when using explicit prior knowledge. Conclusions

Thank you