Applying Bayesian networks to modeling of cell signaling pathways Kathryn Armstrong and Reshma Shetty.

Slides:



Advertisements
Similar presentations
Autonomic Scaling of Cloud Computing Resources
Advertisements

A Tutorial on Learning with Bayesian Networks
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 27, 2012.
Probabilistic modelling in computational biology Dirk Husmeier Biomathematics & Statistics Scotland.
Computational Modeling of the Cell Cycle
Mechanistic models and machine learning methods for TIMET Dirk Husmeier.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
Review: Bayesian learning and inference
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
Regulatory Network (Part II) 11/05/07. Methods Linear –PCA (Raychaudhuri et al. 2000) –NIR (Gardner et al. 2003) Nonlinear –Bayesian network (Friedman.
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
Bayes Nets Rong Jin. Hidden Markov Model  Inferring from observations (o i ) to hidden variables (q i )  This is a general framework for representing.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
Data Mining CS 341, Spring 2007 Lecture 4: Data Mining Techniques (I)
CSE 590ST Statistical Methods in Computer Science Instructor: Pedro Domingos.
Computational Models in Systems Biology Karan Mangla 22 nd April, 2008.
CIS 410/510 Probabilistic Methods for Artificial Intelligence Instructor: Daniel Lowd.
Bayesian networks More commonly called graphical models A way to depict conditional independence relationships between random variables A compact specification.
CSE 515 Statistical Methods in Computer Science Instructor: Pedro Domingos.
Construction of large signaling pathways using an adaptive perturbation approach with phosphoproteomic data Ioannis N. Melas, Alexander Mitsos, Dimitris.
S ystems Analysis Laboratory Helsinki University of Technology Games and Bayesian Networks in Air Combat Simulation Analysis M.Sc. Jirka Poropudas and.
Bayesian network models of Biological signaling pathways
Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks Dirk Husmeier Adriano V. Werhli.
Statistical Bioinformatics QTL mapping Analysis of DNA sequence alignments Postgenomic data integration Systems biology.
CSCE555 Bioinformatics Lecture 16 Identifying Differentially Expressed Genes from microarray data Meeting: MW 4:00PM-5:15PM SWGN2A21 Instructor: Dr. Jianjun.
Reverse engineering gene regulatory networks Dirk Husmeier Adriano Werhli Marco Grzegorczyk.
Learning regulatory networks from postgenomic data and prior knowledge Dirk Husmeier 1) Biomathematics & Statistics Scotland 2) Centre for Systems Biology.
Combined Experimental and Computational Modeling Studies at the Example of ErbB Family Birgit Schoeberl.
Aprendizagem Computacional Gladys Castillo, UA Bayesian Networks Classifiers Gladys Castillo University of Aveiro.
Bayesian Networks for Data Mining David Heckerman Microsoft Research (Data Mining and Knowledge Discovery 1, (1997))
Bayesian Statistics and Belief Networks. Overview Book: Ch 13,14 Refresher on Probability Bayesian classifiers Belief Networks / Bayesian Networks.
Anton Supercomputer Brandon Dean 4/28/15. History Named after Antonie van Leeuwenhoek – “father of microbiology” Molecular Dynamics (MD) simulations were.
Computing & Information Sciences Kansas State University Wednesday, 22 Oct 2008CIS 530 / 730: Artificial Intelligence Lecture 22 of 42 Wednesday, 22 October.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 28 of 41 Friday, 22 October.
Graphical Rule-Based Modeling of Signal-Transduction Systems
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Randomized Algorithms for Bayesian Hierarchical Clustering
Problem Limited number of experimental replications. Postgenomic data intrinsically noisy. Poor network reconstruction.
Computing & Information Sciences Kansas State University Data Sciences Summer Institute Multimodal Information Access and Synthesis Learning and Reasoning.
Inferring gene regulatory networks with non-stationary dynamic Bayesian networks Dirk Husmeier Frank Dondelinger Sophie Lebre Biomathematics & Statistics.
Learning Bayesian networks from postgenomic data with an improved structure MCMC sampling scheme Dirk Husmeier Marco Grzegorczyk 1) Biomathematics & Statistics.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
Artificial Intelligence Research Laboratory Bioinformatics and Computational Biology Program Computational Intelligence, Learning, and Discovery Program.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
CITS7212: Computational Intelligence An Overview of Core CI Technologies Lyndon While.
Boolean Networks and Biology Peter Lee Shaun Lippow BE.400 Final Project December 10, 2002.
Marc Fink & Yan Liu & Shangying Wang Student Project Proposal
To accompany Quantitative Analysis for Management, 7e by (Render/Stair 4-1 © 2000 by Prentice Hall, Inc., Upper Saddle River, N.J Quantitative Analysis.
Learning disjunctions in Geronimo’s regression trees Felix Sanchez Garcia supervised by Prof. Dana Pe’er.
COV outline Case study 1: Model with expert judgment (limited data): Perceived probability of explosive magmatic eruption of La Soufriere, Guadeloupe,
Mechanistic models and machine learning methods for TIMET
Computational methods for inferring cellular networks II Stat 877 Apr 17 th, 2014 Sushmita Roy.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Data Analysis Patrice Koehl Department of Biological Sciences
CS 2750: Machine Learning Directed Graphical Models
Dynamic Bayesian Networks
Statistics for Psychology
CSE 515 Statistical Methods in Computer Science
Nat. Rev. Endocrinol. doi: /nrendo
Polydactyly L.A. Lettice, et al. Proc. Natl. Acad. Sci. USA 2002; 99(11):
Bayesian Statistics and Belief Networks
CSCI2950-C Lecture 13 Network Motifs; Network Integration
Improving Boolean Networks to Model Signaling Pathways
Network Inference Chris Holmes Oxford Centre for Gene Function, &,
Prepared by: Mahmoud Rafeek Al-Farra
A and B, ROC curves of the proteomics panel diagnosis.
Presentation transcript:

Applying Bayesian networks to modeling of cell signaling pathways Kathryn Armstrong and Reshma Shetty

Outline Biological model system (MAPK) Overview of Bayesian networks Design and development Verification Correlation with experimental data Issues Future work

MAPK Pathway K-PP KK-PP KKK*KKK E1E2 KKKK-P KK-PKK’ase K’ase

Overview of Bayesian Networks BurglaryEarthquake Alarm P(A)P(^A)BE No YesNo NoYes Yes Givens:

Bayesian network model K-PP KK-PP KKK*KKK E1E2 KKKK-P KK-PKK’ase K’ase

Normalized concentrations of all species Discretized continuous concentration curves at 20 states Considered steady-state behavior Simplifying Assumptions

The key factor in determining the performance of a Bayesian network is the data used to train the network. Training data Probability tables Bayesian network

Network training I: Data source Current experimental data sets were not sufficient to provide enough information Relied on ODE model to generate training set (Huang et al.) Captured the essential steady-state behavior of the MAPK signaling pathway

Network training II: Poor data variation

Network training III: incomplete versus complete data sets 4D Time = (# samples) 4 E1 1D x 4 E2 MAPKPase MAPKKPase Time = (# samples) x 4

Verification: P(Kinase | E1, P’ases) Huang et al. Bayesian network

Verification: P(E1 | MAPK-PP, P’ases)

C.F. Huang and J.E. Ferrell, Proc. Natl. Acad. Sci. USA 93, (1996). Correlation with experimental data

J.E. Ferrell and E.M. Machleder, Science 280, 895 (1998).

Where does our Bayesian network fail?

Inference from incomplete data K-PP KK-PP KKK*KKK E1E2 KKKK-P KK-PKK’ase K’ase

Future work Time incorporation to represent signaling dynamics Continuous or more finely discretized sampling and modeling of node values Priors Bayesian posterior Structure learning

Open areas of research Should steady state behavior be modeled with a directed acyclic graph? Cyclic networks Hard, but doable Theoretically impossible Need an alternate way to represent feedback loops

Why use a Bayesian network? ODE’s require detailed kinetic and mechanistic information on the pathway. Bayesian networks can model pathways well when large amounts of data are available regardless of how well the pathway is understood.

Acknowledgments Kevin Murphy Doug Lauffenburger Paul Matsudaira Ali Khademhosseini BE400 students

References A.R. Asthagiri and D.A. Lauffenburger, Biotechnol. Prog. 17, 227 (2001). A.R. Asthagiri, C.M. Nelson, A.F. Horowitz and D.A. Lauffenburger, J. Biol. Chem. 274, (1999). J.E. Ferrell and R.R. Bhatt, J. Biol. Chem. 272, (1997). J.E. Ferrell and E.M. Machleder, Science 280, 895 (1998). C.F. Huang and J.E. Ferrell, Proc. Natl. Acad. Sci. USA 93, (1996). F. V. Jensen. Bayesian Networks and Decision Graphs. Springer: New York, K.A. Gallo and G.L. Johnson, Nat. Rev. Mol. Cell Biol. 3, 663 (2002). K.P. Murphy, Computing Science and Statistics. (2001). S. Russell and P. Norvig. Artificial Intelligence: A Modern Approach. Prentice Hall: New York, K Sachs, D. Gifford, T. Jaakkola, P. Sorger and D.A. Lauffenburger, Science STKE 148, 38 (2002).

Network training IV: final data set E1E2 (P’ase)MAPKKPaseMAPKPaseMAPK-PP

Network training V: Final concentration ranges

Network training III: Observation of all input combinations E1 MAPKKPase E2 4D Visualization 3D Visualization 2D Visualization Time = (# samples) 4 1D Visualization E2 MAPKPase MAPKKPase