Reverse Engineering of Genetic Networks (Final presentation)

Slides:



Advertisements
Similar presentations
A Tutorial on Learning with Bayesian Networks
Advertisements

DREAM4 Puzzle – inferring network structure from microarray data Qiong Cheng.
1 Some Comments on Sebastiani et al Nature Genetics 37(4)2005.
Probabilistic modelling in computational biology Dirk Husmeier Biomathematics & Statistics Scotland.
Bayesian Estimation in MARK
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
Structural Inference of Hierarchies in Networks BY Yu Shuzhi 27, Mar 2014.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Bayesian Reasoning: Markov Chain Monte Carlo
Industrial Engineering College of Engineering Bayesian Kernel Methods for Binary Classification and Online Learning Problems Theodore Trafalis Workshop.
Clustering short time series gene expression data Jason Ernst, Gerard J. Nau and Ziv Bar-Joseph BIOINFORMATICS, vol
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Mutual Information Mathematical Biology Seminar
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Particle filters (continued…). Recall Particle filters –Track state sequence x i given the measurements ( y 0, y 1, …., y i ) –Non-linear dynamics –Non-linear.
Course overview Tuesday lecture –Those not presenting turn in short review of a paper using the method being discussed Thursday computer lab –Turn in short.
CENTER FOR BIOLOGICAL SEQUENCE ANALYSIS Bayesian Inference Anders Gorm Pedersen Molecular Evolution Group Center for Biological Sequence Analysis Technical.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1 Bayesian Networks Chapter ; 14.4 CS 63 Adapted from slides by Tim Finin and Marie desJardins. Some material borrowed from Lise Getoor.
Cristina Manfredotti D.I.S.Co. Università di Milano - Bicocca An Introduction to the Use of Bayesian Network to Analyze Gene Expression Data Cristina Manfredotti.
Super-Resolution of Remotely-Sensed Images Using a Learning-Based Approach Isabelle Bégin and Frank P. Ferrie Abstract Super-resolution addresses the problem.
Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks Dirk Husmeier Adriano V. Werhli.
A Brief Introduction to Graphical Models
WSEAS AIKED, Cambridge, Feature Importance in Bayesian Assessment of Newborn Brain Maturity from EEG Livia Jakaite, Vitaly Schetinin and Carsten.
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Learning Structure in Bayes Nets (Typically also learn CPTs here) Given the set of random variables (features), the space of all possible networks.
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
Introduction to MCMC and BUGS. Computational problems More parameters -> even more parameter combinations Exact computation and grid approximation become.
Image Segmentation Seminar III Xiaofeng Fan. Today ’ s Presentation Problem Definition Problem Definition Approach Approach Segmentation Methods Segmentation.
Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Reverse engineering gene regulatory networks Dirk Husmeier Adriano Werhli Marco Grzegorczyk.
Learning regulatory networks from postgenomic data and prior knowledge Dirk Husmeier 1) Biomathematics & Statistics Scotland 2) Centre for Systems Biology.
Unsupervised Learning: Clustering Some material adapted from slides by Andrew Moore, CMU. Visit for
Markov Random Fields Probabilistic Models for Images
Learning Linear Causal Models Oksana Kohutyuk ComS 673 Spring 2005 Department of Computer Science Iowa State University.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Randomized Algorithms for Bayesian Hierarchical Clustering
Learning the Structure of Related Tasks Presented by Lihan He Machine Learning Reading Group Duke University 02/03/2006 A. Niculescu-Mizil, R. Caruana.
Molecular Systematics
Learning With Bayesian Networks Markus Kalisch ETH Zürich.
1 CMSC 671 Fall 2001 Class #21 – Tuesday, November 13.
Inferring gene regulatory networks with non-stationary dynamic Bayesian networks Dirk Husmeier Frank Dondelinger Sophie Lebre Biomathematics & Statistics.
Tracking Multiple Cells By Correspondence Resolution In A Sequential Bayesian Framework Nilanjan Ray Gang Dong Scott T. Acton C.L. Brown Department of.
Slides for “Data Mining” by I. H. Witten and E. Frank.
Learning Bayesian networks from postgenomic data with an improved structure MCMC sampling scheme Dirk Husmeier Marco Grzegorczyk 1) Biomathematics & Statistics.
Lecture 2: Statistical learning primer for biologists
An Introduction to Markov Chain Monte Carlo Teg Grenager July 1, 2004.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
04/21/2005 CS673 1 Being Bayesian About Network Structure A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller.
Bayesian statistics named after the Reverend Mr Bayes based on the concept that you can estimate the statistical properties of a system after measuting.
Computational methods for inferring cellular networks II Stat 877 Apr 17 th, 2014 Sushmita Roy.
HW7: Evolutionarily conserved segments ENCODE region 009 (beta-globin locus) Multiple alignment of human, dog, and mouse 2 states: neutral (fast-evolving),
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Bayesian Networks Chapter 2 (Duda et al.) – Section 2.11 CS479/679 Pattern Recognition Dr. George Bebis.
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
MCMC Output & Metropolis-Hastings Algorithm Part I
Incorporating graph priors in Bayesian networks
Advanced Statistical Computing Fall 2016
Markov Networks.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Class #19 – Tuesday, November 3
Markov Networks.
Presentation transcript:

Reverse Engineering of Genetic Networks (Final presentation) Ji Won Yoon (s0344084) supervised by Dr. Dirk Husmeier. MSc in Informatics at Edinburgh University, J.Yoon@sms.ed.ac.uk

Reverse Engineering What is reverse engineering of gene network? Missing gene + “up” and “down” data from micro array - Relevance Network My own method - MCMC for Bayesian network

Past works Comparison of existing approaches to the reverse engineering of genetic networks, Mutual information relevance networks My own method Bayesian networks using Markov Chain Monte Carlo method Applying all methods to synthetic data generated from a gene network simulator. Applying to Biological data Diffuse large B cell Lymphoma gene expression data Arabidopsis gene expression data

Relevance Network (Butte, 2000) Using mutual information MI(A, B) = H(A) – H(A|B) = H(B) – H(B|A) = MI(B, A) -> Symmetric MI(A, B) = H(A) + H(B) – H(A, B) Mutual information is zero if two genes are independent. Pair wise relation

Relevance Network (cont.) Relevance Network (Butte, 2000) Useful only to local relation due to pair wise relation. Important to select proper threshold to get good relations. Bootstrapping (Comparison of results in real data and in randomly permuted data) Difficulty to identify the relation with two parents due to the locality. MI(A, [B, C, D])> MI(A, B)< MI(A, C)< MI(A, D)< Cannot detect XOR operation MI(A, C) = MI(B, C) = 0 No direction of edges due to symmetric property Fast and light computation. Useful for a number of genes C D A B A B C 1

My method (using mutual information) Based on Scale free network Crucial genes will have more connections than other genes. A B C D A B C D On insert new gene F, A will have more chance to have it than other genes. E G = (N, E) = (N, E, L) (L is level information}

My method (Insertion step) Finding better parents and merging Clusters Threshold = 0.3 ( ) a 9 1 4 5 6 7 S4 3 8 S1 10 11 2 1 12 5 a 4 7 S3 6 S2 MI(1, a) = 0.34 MI(4, a) = 0.28 MI(5, a) = 0.35 MI(6, a) = 0.31 MI(7, a) = 0.4

My method (Deletion step) Assumption The network generated from insertion step of my method is in stationary state in marginal log likelihood except one edge, which is investigated to check the connection Three case in an edge e X->Y, X<-Y, and X Y P (D | M) = U * P (X | pa (X)) * P (Y | pa (Y))

My method (Deletion step) X Y Graph G e A C D H I B E F

My method Mainly two steps Insertion and deletion steps Insertion step f g h c d a b = 0.5 = 0.4 = 0.3 Continue up to t = 0. … Insertion step Deletion step

My method Advantage Based on Biological facts (Scale Free Network) No need of thresholds Online approach Scalability Easy to explore sub-networks Fast computation

My method Disadvantage Input order dependency Risky in exploring parents in data with big noise values. (It can be over-fitted to training data) 61 % edges are less order dependent (in part B)

Bayesian network with MCMC D E Problem 1 Problem 2 Left: in large data set. Right: in small data set.

Bayesian network with MCMC MCMC (Markov Chain Monte Carlo) Inference rule for Bayesian Network Sample from the posterior distribution Proposal Move : Given M_old, propose a new network M_new with probability Acceptance and Rejection :

Bayesian network with MCMC

MCMC in Bayes Net toolbox Hasting factor The proposal probability is calculated from the number of neighbours of the model.

Improvement of MCMCs Fan-in The sparse data leads the prior probability to have a non-negligible influence on the posterior P(M|D). Limit the maximum number of edges converging on a node, fan-in. If FI(M) > a, P(M)=0. Otherwise, P(M)=1. The time complexity reduced largely A B C D E Acceptable configuration of child and parents in fan-in 3

Improvement of MCMCs DAG to CPDAG X Y X Y (DAG : Directed Acyclic Graph, CPDAG : Completed Partially Directed Acyclic Graph) X Y X Y P(X, Y) = P(X)P(X|Y) = P(Y|P(Y|X) Set of all equivalent DAGs DAG to CPDAG DE is reversible others are compelled.

Improvement of MCMCs This CPDAG concept bring several advantages: The space of equivalent classes is more reduced. It is easy to trap in local optimum in moving DAG spaces. Incorporating CPDAG to MCMC

MCMCMC Trapping A A : global optima B : local optima B Easy to be trapped in local optima B.  Multi chains with different temperatures will be useful to escape from it.

MCMCMC Trapping

MCMCMC A super chain, S Acceptance ratios of a super chain

Importance Sampling . Partition function Proposal distribution Acceptance probability We only case the prior distribution for acceptance. Importance Sampling is also combined with MCMCMC. MCMCMC with Importance Sampling : Likelihood for configuration of a node n and its parents

Order MCMC It sample over total orders not over structures. A B C A B A C C A B C B A A B C A C B B C

Order MCMC It sample over total orders not over structures. Proposal move flipping two nodes of the previous order Computational limitations Using candidate sets Sets of parents with the highest scores in likelihood for each node Reduces the computation time.

Order MCMC

Order MCMC Selection features We can extract the edges by approximating and averaging under the stationary distribution, where

Synthetical data 41th to 50th genes are not connected.

Synthetic data - MCMCMC with Importance Sampling has the best performance. - Order MCMC is the second. - Order MCMC is much faster than MCMCMC with Importance Sampling.

Synthetic data I changed one parameters for MCMC simulation. Standard application (using standard parameters) Change a noise value (Decrease noise value to 0.1) 3) Change a training data size (Decrease the size to 50) 4) Change the number of iterations (Increase the number to 50000) Standard parameters ( MCMC in Bayes Net Toolbox ) training data size:200, noise value:0.3, the number of iterations: 5000 (5000 samples and 5000 burn-ins)

Synthetic data

Synthetic data 2 1 Convergence MCMC in BNT MCMCMC Importance Sampling (IM) 3) MCMCMC Importance Sampling (ID) 4) Order MCMC 3 4 training set size : 200 noise : 0.3 5000 iterations.

Synthetic data MCMCMC (Burn-in# + Sample #) Left: 5000 + 5000 Right: 100000 + 100000 Acceptance ratios Left: MCMC in BNT, Right : Order MCMC Middle: MCMCMC with Importance Sampling

Diffuse large B cell lymphoma Data Data discretisation I used K-means algorithms to discretise gene expression levels for each genes since the stationary level for each gene can be different from others. (up, down and normal) Problem of this discretisation If there are too many noises, the noises can make fluctuations Finally, this method can not work well for gene3.

Diffuse large B cell lymphoma Data Comparison of convergence MCMC in BNT MCMCMC with Importance Sampling(ID) Order MCMC # of genes : 27 Training data size : 105 Iterations : 20000

Diffuse large B cell lymphoma Data Comparison of Acceptance Ratios The number of genes : 27, Training data size : 105, Iterations : 20000

Gene expression inoculated by viruses in susceptible Arabidopsis thaliana plants 4) Viruses Cucumber mosaic cucumovirus Oil seed rape tobamovirus Turnip vein clearing tobamovirus Potato virus X potexvirus Turnip mosaic potyvirus 1) 5) 2) 3) Inoculation DAI = Day after inoculation 1DAI 2DAI 3DAI 4DAI 5DAI 7DAI Symptom occurs. Gene a Training data : 127 genes with 20 data size ( 4 DAIs * 5 viruses )

Gene expression inoculated by viruses in susceptible Arabidopsis thaliana plants only for 20 genes (1DAI and 2DAI) 10000 samples from MCMCMC with Importance Sampling(ID) 1000 samples from my method

Gene expression inoculated by viruses in susceptible Arabidopsis thaliana plants 127 genes Average global connectivity = 1.5847 Genes with Higher connectivity

Gene expression inoculated by viruses in susceptible Arabidopsis thaliana plants 127 genes p-value check for transcription function - f is the number of genes with j th function in 127 genes. - m is the number of genes with j th function in 14 genes.

Gene expression inoculated by viruses in susceptible Arabidopsis thaliana plants for 127 genes from my method (100 samples)

Conclusion We need to select methodologies depending on the characteristics of training data. To obtain the closest result to real networks, MCMCMC with Importance and Order MCMC are suitable. MCMCMC with Importance Sampling has the best performance but it is slower than other MCMCs. Order MCMC has the second performance but it is four times faster than MCMCMC with Importance Sampling. If we want to process large scale data and we do not have enough time to run MCMCs, Relevance Network and My method are proper. Also, several methods generate different networks so that combining them will give better results.

Conclusion Biological meaning Transcription genes have higher connectivities more than other genes (from my method). That is, genes with transcription function may act as hubs in a network for response against viruses in Arabidopsis thaliana plant.