Learning Module Networks Eran Segal Stanford University Joint work with: Dana Pe’er (Hebrew U.) Daphne Koller (Stanford) Aviv Regev (Harvard) Nir Friedman.

Slides:



Advertisements
Similar presentations
Learning Probabilistic Relational Models Daphne Koller Stanford University Nir Friedman Hebrew University Lise.
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Inferring Quantitative Models of Regulatory Networks From Expression Data Iftach Nachman Hebrew University Aviv Regev Harvard Nir Friedman Hebrew University.
. Inferring Subnetworks from Perturbed Expression Profiles D. Pe’er A. Regev G. Elidan N. Friedman.
Introduction of Probabilistic Reasoning and Bayesian Networks
Rich Probabilistic Models for Gene Expression Eran Segal (Stanford) Ben Taskar (Stanford) Audrey Gasch (Berkeley) Nir Friedman (Hebrew University) Daphne.
From Sequence to Expression: A Probabilistic Framework Eran Segal (Stanford) Joint work with: Yoseph Barash (Hebrew U.) Itamar Simon (Whitehead Inst.)
Nir Friedman, Iftach Nachman, and Dana Peer Announcer: Kyu-Baek Hwang
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
. Class 1: Introduction. The Tree of Life Source: Alberts et al.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Cs726 Modeling regulatory networks in cells using Bayesian networks Golan Yona Department of Computer Science Cornell University.
Goal: Reconstruct Cellular Networks Biocarta. Conditions Genes.
Recovering Articulated Object Models from 3D Range Data Dragomir Anguelov Daphne Koller Hoi-Cheung Pang Praveen Srinivasan Sebastian Thrun Computer Science.
6. Gene Regulatory Networks
Module Networks Discovering Regulatory Modules and their Condition Specific Regulators from Gene Expression Data Cohen Jony.
Copyright N. Friedman, M. Ninio. I. Pe’er, and T. Pupko. 2001RECOMB, April 2001 Structural EM for Phylogentic Inference Nir Friedman Computer Science &
Artificial Intelligence Term Project #3 Kyu-Baek Hwang Biointelligence Lab School of Computer Science and Engineering Seoul National University
Introduction to molecular networks Sushmita Roy BMI/CS 576 Nov 6 th, 2014.
Causal Models, Learning Algorithms and their Application to Performance Modeling Jan Lemeire Parallel Systems lab November 15 th 2006.
Cristina Manfredotti D.I.S.Co. Università di Milano - Bicocca An Introduction to the Use of Bayesian Network to Analyze Gene Expression Data Cristina Manfredotti.
Inferring subnetworks from perturbed expression profiles Dana Pe’er, Aviv Regev, Gal Elidan and Nir Friedman Bioinformatics, Vol.17 Suppl
Inferring Cellular Networks Using Probabilistic Graphical Models Jianlin Cheng, PhD University of Missouri 2009.
Bayes Net Perspectives on Causation and Causal Inference
Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks Dirk Husmeier Adriano V. Werhli.
Whole Genome Expression Analysis
MATISSE - Modular Analysis for Topology of Interactions and Similarity SEts Igor Ulitsky and Ron Shamir Identification.
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
12/07/2008UAI 2008 Cumulative Distribution Networks and the Derivative-Sum-Product Algorithm Jim C. Huang and Brendan J. Frey Probabilistic and Statistical.
Using Bayesian Networks to Analyze Expression Data By Friedman Nir, Linial Michal, Nachman Iftach, Pe'er Dana (2000) Presented by Nikolaos Aravanis Lysimachos.
Outline Who regulates whom and when? Model Learning algorithm Evaluation Wet lab experiments Perspective: why does it work? Reg. ACGTGC.
Data Analysis with Bayesian Networks: A Bootstrap Approach Nir Friedman, Moises Goldszmidt, and Abraham Wyner, UAI99.
Reverse engineering gene regulatory networks Dirk Husmeier Adriano Werhli Marco Grzegorczyk.
Using Bayesian Networks to Analyze Whole-Genome Expression Data Nir Friedman Iftach Nachman Dana Pe’er Institute of Computer Science, The Hebrew University.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Unsupervised Learning: Clustering Some material adapted from slides by Andrew Moore, CMU. Visit for
Lectures 2 – Oct 3, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall.
CS5263 Bioinformatics Lecture 20 Practical issues in motif finding Final project.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Lectures 9 – Oct 26, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall.
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Problem Limited number of experimental replications. Postgenomic data intrinsically noisy. Poor network reconstruction.
Slides for “Data Mining” by I. H. Witten and E. Frank.
By: Amira Djebbari and John Quackenbush BMC Systems Biology 2008, 2: 57 Presented by: Garron Wright April 20, 2009 CSCE 582.
Class 23, 2001 CBCl/AI MIT Bioinformatics Applications and Feature Selection for SVMs S. Mukherjee.
MCMC in structure space MCMC in order space.
Dependency networks Sushmita Roy BMI/CS 576 Nov 25 th, 2014.
Introduction to biological molecular networks
Advances in Bayesian Learning Learning and Inference in Bayesian Networks Irina Rish IBM T.J.Watson Research Center
Module Networks BMI/CS 576 Mark Craven December 2007.
Nonlinear differential equation model for quantification of transcriptional regulation applied to microarray data of Saccharomyces cerevisiae Vu, T. T.,
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
1 Structure Learning (The Good), The Bad, The Ugly Inference Graphical Models – Carlos Guestrin Carnegie Mellon University October 13 th, 2008 Readings:
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
Computational methods for inferring cellular networks II Stat 877 Apr 17 th, 2014 Sushmita Roy.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Dependency Networks for Inference, Collaborative filtering, and Data Visualization Heckerman et al. Microsoft Research J. of Machine Learning Research.
Inferring Regulatory Networks from Gene Expression Data BMI/CS 776 Mark Craven April 2002.
Learning gene regulatory networks in Arabidopsis thaliana
Irina Rish IBM T.J.Watson Research Center
Recovering Temporally Rewiring Networks: A Model-based Approach
1 Department of Engineering, 2 Department of Mathematics,
A Short Tutorial on Causal Network Modeling and Discovery
1 Department of Engineering, 2 Department of Mathematics,
1 Department of Engineering, 2 Department of Mathematics,
Network Inference Chris Holmes Oxford Centre for Gene Function, &,
Learning Probabilistic Graphical Models Overview Learning Problems.
Quantum-Classical Hybrid Machine Learning for Gene Regulatory Pathways
Presentation transcript:

Learning Module Networks Eran Segal Stanford University Joint work with: Dana Pe’er (Hebrew U.) Daphne Koller (Stanford) Aviv Regev (Harvard) Nir Friedman (Hebrew U.)

Learning Bayesian Networks Density estimation Model data distribution in population Probabilistic inference: Prediction Classification Dependency structure Interactions between variables Causality Scientific discovery Data INTL MSFT MOT NVLS

Stock Market Learn dependency of stock prices as a function of Global influencing factors Sector influencing factors Price of other major stocks Mar.’02May.’02Aug.’02Oct.’02Jan.’03 Jan.’02 MSFT DELL INTL NVLS MOTI MSFT DELL INTL NVLS MOT

Stock Market Learn dependency of stock prices as a function of Global influencing factors Sector influencing factors Price of other major stocks Mar.’02May.’02Aug.’02Oct.’02Jan.’03 Jan.’02 MSFT DELL INTL NVLS MOTI DELL INTL NVLS MOT MSFT

Stock Market Learn dependency of stock prices as a function of Global influencing factors Sector influencing factors Price of other major stocks Mar.’02May.’02Aug.’02Oct.’02Jan.’03 Jan.’02 MSFT DELL INTL NVLS MOTI INTL MSFT DELL NVLS MOT Bayesian Network

Fragment of learned BN Stock Market 4411 stocks (variables) 273 trading days (instances) from Jan.’02 – Mar.’03 Problems Statistical robustness Interpretability

Key Observation Many stocks depend on the same influencing factors in much the same way Example: Intel, Novelus, Motorola, Dell depend on the price of Microsoft Many other domains with similar characteristics Gene expression Collaborative filtering Computer network performance … Mar.’02May.’02Aug.’02Oct.’02Jan.’03 Jan.’02 MSFT DELL INTL NVLS MOTI

INTL MSFT MOT DELL AMAT HPQ CPD 2 CPD 1 CPD 3 Bayesian Network The Module Network Idea CPD 6 CPD 3 CPD 5 CPD 1 CPD 2 CPD 4 INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I Module Network

Problems and Solutions Statistical robustness Interpretability Share parameters and dependencies between variables with similar behavior Explicit modeling of modular structure

Outline Module Network Probabilistic model Learning the model Experimental results

Module Network Components Module Assignment Function A(MSFT)=M I A(MOT)=A(DELL)=A(INTL) =M II A(AMAT)= A(HPQ)=M III INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I INTL MSFT MOT DELL AMAT HPQ

Module Network Components Module Assignment Function Set of parents for each module Pa(M I )=  Pa(M II )={MSFT} Pa(M III )={DELL, INTL} INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I

Module Network Components Module Assignment Function Set of parents for each module CPD template for each module INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I

Ground Bayesian Network A module network induces a ground BN over X A module network defines a coherent probabilty distribution over X if the ground BN is acyclic INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I INTL MSFT MOT DELL AMAT HPQ Ground Bayesian Network

Module Graph Nodes correspond to modules M i  M j if at least one variable in M i is a parent of M j INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I MIMI M II M III Module graph Theorem: The ground BN is acyclic if the module graph is acyclic Acyclicity checked efficiently using the module graph

Outline Module Network Probabilistic model Learning the model Experimental results

Learning Overview Given data D, find assignment function A and structure S that maximize the Bayesian score Marginal data likelihood Data likelihood Parameter prior Marginal likelihood Assignment / structure prior

Instance 3 Likelihood Function Module III Module II Module I INTL MSFT MOT DELL AMAT HPQ Instance 1 Instance 2 MIMI  M II |MSFT  M III |DELL,INTL Sufficient statistics of (X,Y) Likelihood function decomposes by modules

Bayesian Score Decomposition Bayesian score decomposes by modules INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I Delete INTL  Module III Module j variablesModule j parents

Bayesian Score Decomposition Bayesian score decomposes by modules INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I A(MOT)=2  A(MOT)=1

Algorithm Overview Find assignment function A and structure S that maximize the Bayesian score Dependency structure S Improve structure Improve assignments Find initial assignment A Assignment function A

Initial Assignment Function x[1] AMAT MSFTDELL MOT HPQ INTL x[2] x[3] x[4] Variables (stocks) Instances (trading days) Find variables that are similar across instances A(MOT)= M II A(INTL)= M II A(DELL)= M II MSFT MOTHPQ AMAT DELL INTL 123

Algorithm Overview Find assignment function A and structure S that maximize the Bayesian score Dependency structure S Improve structure Improve assignments Find initial assignment A Assignment function A

Learning Dependency Structure Heuristic search with operators Add/delete parent for module Cannot reverse edges Handle acyclicity Can be checked efficiently on the module graph Efficient computation After applying operator for module M j, only update score of operators for module M j INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I MIMI M II M III X INTL  Module I INTL  Module III X  MSFT  Module II

Learning Dependency Structure Structure search done at module level Parent selection Reduced search space relative to BN Acyclicity checking Individual variables only used for computation of sufficient statistics

Algorithm Overview Find assignment function A and structure S that maximize the Bayesian score Dependency structure S Improve structure Improve assignments Find initial assignment A Assignment function A

Learning Assignment Function A(DELL)=M I Score: 0.7 INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I DELL

Learning Assignment Function A(DELL)=M I Score: 0.7 A(DELL)=M II Score: 0.9 INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I DELL

Learning Assignment Function A(DELL)=M I Score: 0.7 A(DELL)=M II Score: 0.9 A(DELL)=M III Score: cyclic! INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I DELL

Learning Assignment Function A(DELL)=M I Score: 0.7 A(DELL)=M II Score: 0.9 A(DELL)=M III Score: cyclic! INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I

Ideal Algorithm Learn the module assignment of all variables simultaneously

Problem Due to acyclicity cannot optimize assignment for variables separately DELL Module I AMAT Module III MSFT Module II HPQ Module IV MIMI M II M III Module Network Module graph M IV A(DELL)=Module IV A(MSFT)=Module III DELL MSFT DELL

Problem Due to acyclicity cannot optimize assignment for variables separately DELL Module I AMAT Module III MSFT Module II HPQ Module IV MIMI M II M III Module Network Module graph M IV A(DELL)=Module IV A(MSFT)=Module III DELL MSFT DELL

Learning Assignment Function Sequential update algorithm Iterate over all variables For each variable, find its optimal assignment given the current assignment to all other variables Efficient computation When changing assignment from M i to M j, only need to recompute score for modules i and j

Learning the Model Initialize module assignment A Optimize structure S Optimize module assignment A For each variable, find its optimal assignment given the current assignment to all other variables INTL MSFT MOT DELL AMAT HPQ Module III Module II Module I INTL MSFT MOT DELL AMAT HPQ MOT

Related Work Bayesian networks Parameter sharing PRMs OOBNs Module Networks Shared structure X  X X  Shared parameters   X   Learn parameter sharing X X  Langseth+al N/A Learn structure X   

Outline Module Network Probabilistic model Learning the model Experimental results Statistical validation Case study: Gene regulation

Learning Algorithm Performance Bayesian score (avg. per gene) Algorithm iterations Structure change iterations

Test data likelihood (per instance) Number of modules 25 instances 50 instances 100 instances 200 instances 500 instances Generalization to Test Data Synthetic data: 10 modules, 500 variables Best performance achieved for models with 10 modules

Generalization to Test Data Test data likelihood (per instance) Number of modules Synthetic data: 10 modules, 500 variables 25 instances 50 instances 100 instances 200 instances 500 instances Gain beyond 100 instances is small

Structure Recovery Graph Synthetic data: 10 modules, 500 variables Number of modules Recovered structure (% correct) 25 instances 50 instances 200 instances 500 instances 100 instances 74% of 2250 parent- child relationships recovered

Stock Market 4411 variables (stocks), 273 instances (trading days) Comparison to Bayesian networks (cross validation) Test Data Log-Likelihood (gain per instance) Number of modules Bayesian network performance

Regulatory Networks Learn structure of regulatory networks: Which genes are regulated by each regulator

Gene Expression Data Measures mRNA level for all genes in one condition Learn dependency of the expression of genes as a function of expression of regulators Experiments Genes Induced Repressed

Gene Expression 2355 variables (genes), 173 instances (arrays) Comparison to Bayesian networks Test Data Log-Likelihood (gain per instance) Number of modules Bayesian network performance

Biological Evaluation Find sets of co-regulated genes (regulatory module) Find the regulators of each module Segal et al., Nature Genetics, /50 30/50

Experimental Design Hypothesis: Regulator ‘X’ activates process ‘Y’ Experiment: Knock out ‘X’ and repeat experiment HAP4  Ypl230W true false true false X ? Segal et al., Nature Genetics, 2003

wt   Ypl230w (hrs.) >16x 341 differentially expressed genes wt  (min.)  Ppt1 >4x wt  (min.)  Kin82 >4x 281 Differentially Expressed Genes Segal et al., Nature Genetics, 2003

Were the differentially expressed genes predicted as targets? Rank modules by enrichment for diff. expressed genes #ModuleSignificance 14 Ribosomal and phosphate metabolism 8/32, 9e3 11 Amino acid and purine metabolism11/53, 1e2 15 mRNA, rRNA and tRNA processing 9/43, 2e2 39 Protein folding6/23, 2e2 30 Cell cycle7/30, 2e2  Ppt1 # ModuleSignificance 39Protein folding7/23, 1e-4 29Cell differentiation6/41, 2e-2 5 Glycolysis and folding5/37, 4e-2 34Mitochondrial and protein fate 5/37, 4e-2  Ypl230w #ModuleSignificance 3Energy and osmotic stress I8/31, 1e4 2Energy, osmolarity & cAMP signaling9/64, 6e3 15 mRNA, rRNA and tRNA processing6/43, 2e2  Kin82 Biological Experiments Validation All regulators regulate predicted modules Segal et al., Nature Genetics, 2003

Summary Probabilistic model for learning modules of variables and their structural dependencies Improved performance over Bayesian networks Statistical robustness Interpretability Application to gene regulation Reconstruction of many known regulatory modules Prediction of targets for unknown regulators