Exploring Network Inference Models Math-in-Industry Camp & Workshop: Michael Grigsby: Cal Poly, Pomona Mustafa Kesir: Northeastern University Nancy Rodriguez:

Slides:



Advertisements
Similar presentations
Pretty-Good Tomography Scott Aaronson MIT. Theres a problem… To do tomography on an entangled state of n qubits, we need exp(n) measurements Does this.
Advertisements

Artificial Intelligence Presentation
PARAMETER ESTIMATION FOR ODES USING A CROSS-ENTROPY APPROACH Wayne Enright Bo Wang University of Toronto.
Lect.3 Modeling in The Time Domain Basil Hamed
Brief introduction on Logistic Regression
CGeMM – University of Louisville Mining gene-gene interactions from microarray data - Coefficient of Determination Marcel Brun – CGeMM - UofL.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Artificial Intelligence in Game Design Introduction to Learning.
Chapter 19 Confidence Intervals for Proportions.
The loss function, the normal equation,
Planning under Uncertainty
Date:2011/06/08 吳昕澧 BOA: The Bayesian Optimization Algorithm.
Maximum likelihood (ML) and likelihood ratio (LR) test
CSCE Review—Fortran. CSCE Review—I/O Patterns: Read until a sentinel value is found Read n, then read n things Read until EOF encountered.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Computational Methods for Management and Economics Carla Gomes Module 3 OR Modeling Approach.
Chapter 4 Multiple Regression.
Maximum likelihood (ML) and likelihood ratio (LR) test
The Basics of Regression continued
Gene Regulatory Networks - the Boolean Approach Andrey Zhdanov Based on the papers by Tatsuya Akutsu et al and others.
Development of Empirical Models From Process Data
Lecture 17 Interaction Plots Simple Linear Regression (Chapter ) Homework 4 due Friday. JMP instructions for question are actually for.
Maximum likelihood (ML)
Introduction to Simulated Annealing 22c:145 Simulated Annealing  Motivated by the physical annealing process  Material is heated and slowly cooled.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Component Reliability Analysis
0 Pattern Classification, Chapter 3 0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda,
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Chapter 8 Introduction to Hypothesis Testing
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Probabilistic Robotics Bayes Filter Implementations.
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
ENM 503 Lesson 1 – Methods and Models The why’s, how’s, and what’s of mathematical modeling A model is a representation in mathematical terms of some real.
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Mobile Robot Localization (ch. 7)
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
Inferring gene regulatory networks with non-stationary dynamic Bayesian networks Dirk Husmeier Frank Dondelinger Sophie Lebre Biomathematics & Statistics.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
De novo discovery of mutated driver pathways in cancer Discussion leader: Matthew Bernstein Scribe: Kun-Chieh Wang Computational Network Biology BMI 826/Computer.
Statistics What is the probability that 7 heads will be observed in 10 tosses of a fair coin? This is a ________ problem. Have probabilities on a fundamental.
Learning Chaotic Dynamics from Time Series Data A Recurrent Support Vector Machine Approach Vinay Varadan.
Nonlinear differential equation model for quantification of transcriptional regulation applied to microarray data of Saccharomyces cerevisiae Vu, T. T.,
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Artificial Intelligence in Game Design Lecture 20: Hill Climbing and N-Grams.
Computacion Inteligente Least-Square Methods for System Identification.
Introduction to emulators Tony O’Hagan University of Sheffield.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Bayesian Neural Networks
Journal club Jun , Zhen.
12. Principles of Parameter Estimation
Ch3: Model Building through Regression
Chapter 3 Component Reliability Analysis of Structures.
Simple Linear Regression - Introduction
1 Department of Engineering, 2 Department of Mathematics,
Effective Social Network Quarantine with Minimal Isolation Costs
Objective of This Course
Regression Models - Introduction
1 Department of Engineering, 2 Department of Mathematics,
Estimating Networks With Jumps
1 Department of Engineering, 2 Department of Mathematics,
The loss function, the normal equation,
Mathematical Foundations of BME Reza Shadmehr
Lecture # 2 MATHEMATICAL STATISTICS
12. Principles of Parameter Estimation
Kalman Filter: Bayes Interpretation
Regression Models - Introduction
Presentation transcript:

Exploring Network Inference Models Math-in-Industry Camp & Workshop: Michael Grigsby: Cal Poly, Pomona Mustafa Kesir: Northeastern University Nancy Rodriguez: University of California, Los Angeles Man Vu: Cal State University, Long Beach

Introduction-Problem Statement Problem proposed by Ruye Wang from Harvey Mudd College. Some biological processes are modeled by networks comprised of a group of interacting components such as genes in a gene regulatory network, neurons in the brain, or proteins. Biologists want to know how the components of the network are related and how they interact to make predictions about the behavior of a biological system.

Introduction Network Inference is an approach for modeling and analyzing networks composed of many interacting component units. That is, given a set of genes a biologist performs a series of experiments to test how the genes affect (excite or inhibit) one another and also determine the magnitude of that affect.

Introduction There are several different mathematical models for network inference each with it’s own advantages and disadvantages. One is the Boolean Network Model which simulates the components by a group of binary nodes interacting with each other that follow logical operations.

Introduction Another is the Linear and Quasi-Linear models that assume the components are linearly or quasi-linearly related in the network. Then there is the Differential Equation (DE) model that simulates the dynamics of the network by a system of differential equations. This is the model we studied.

Introduction Given a set of n nodes (genes) in the network and a set of k data points taken over time the differential equation governing the dynamics of the network is: r i v i ’ (t) + λ i v i (t) = g[Σ T im v m (t) + h i ] Where (i=1,…,n and t=1,…m) v i (t) is the observed data and the other parameters are unknown where r i is a time constant, λ i is a scaling factor, T im is a constant that describes how node m affects node i, and h i is a constant.

Introduction The goal is to find estimates for the n by n matrix T, along with the other unknown parameters from the observed data V i (t). However an optimal search of a O(n 2 )- dimensional space must be conducted in order to find the parameters that minimize the error. This is very computationally expensive and can realistically be done for a network with only a small number of nodes.

Our first Attempt rv’ + λv= g(X) Try to find a relationship between r and λ to reduce parameters. λ r λ r = f(λ)

But How? Given g(x), some s function with range (-1,1) g(x)= e x -1 then g’(x) Є (0,1] e x +1 rv’ + λv = g(X) rv’’ + λv = g’(X) α Є (-1,1) and β Є (0,1] a b c d rλrλ = αβαβ

Other Methods Most existing methods requires a heuristic approach Requires many assumptions and parallel programming Other than heuristic methods, statistical methods are viable but not feasible for large numbers of genes

Bayesian Networks Statistical approach for modeling gene networks Treats each gene as a random variable Joint distribution over all genes represents the cell states Goal: estimate and study the structures of the distributions

To Name a Few Boolean Networks: uses 0’s and 1’s to represent the excitation or not Differential Equation Models:  Many unknown parameters and assumptions  Nonlinear models needs to be linearized  Computationally costly for large number of genes

Simulated Annealing 1. Let X := initial configuration 2. Let E := Energy(X) 3. Let i = random move from the Moveset 4. Let Ei := Eval(move(X,i)) 5. If E < Ei then X := move(X,i) E := Ei Else with some probability, accept the move even though things get worse: X := move(X,i) E := Ei 6. Goto 3 unless we have reached t_max Allowable moves. Choosing this is key!

Algorithm: Choosing (τ,λ) The domain of g -1 is (-1,1)! This is where conditions for λ come in.

Algorithm: Solving for T and h i.e

Algorithm: Decreasing Cost Tm decreases with each iteration. The more iterations the less likely you make possible “bad moves” same for change in cost.

Possible Area of Improvement If we had more time where would we focus?  Simulated Annealing is a good idea provided you move within your moveset intelligently.  Choosing the moveset is also important, for us g(x) helps restrict the domain of λ based on τ. How do you know the domain of τ.  Finding the derivative matrix can possibly be improved.  Recovering the data, solving the ODE.  Choosing the correct energy function.  Solving the system of algebraic equations.

Ideas for moving within Moveset Recall the computations: Might be better to check if λ 0 lies within the range dictated by τ 1, and compare C(λ 0, τ 1 ) to C(λ 0, τ 0 ). Neighborhood of search must be small enough.

When k is not big enough, i.e. when k<n; One obvious way could be:  Once we interpolate to get v i (t);  We can get as many time observations as we need, i.e. we can make k as big as necessary.

Another way could be: Again taking DE as the model;  We can reduce the number of nodes, i.e. get a smaller number of nodes  To get all unknowns, ,Tij, hi we need to have k=n+1 or bigger. If k<n, then eliminate (n-k-1) nodes.  It can result in a loss of important data, the way we do that is really important. Thinking of v i (t) ’s as functions, it’s possible that all n of them are linearly independent.

Functional Data Analysis (FDA) (*) could be extremely helpful in this manner. The thing is, in biological applications, we usually have huge n(~10000), and FDA is extremely useful in dealing with big data samples. (*) Ramsay, J. O. and Silverman, B.W. (2002) Applied functional data analysis : methods and case studies, Springer series in statistics, New York ; London : Springer (*) Ramsay, J. O. and Silverman, B.W. (2005) Functional data analysis, 2nd ed., New York : Springer Also available to view online through Claremont campus:

Working with the DE model, one immediately notices that computational cost (O(n 2 )) is a major obstacle. As long as complexity of FDA is not as big as O(n 2 ), at does not make things any worse. (Actually, even if O(n 2 ) is fine).