Lecture 3: Markov processes, master equation

Slides:



Advertisements
Similar presentations
Scientific Computing QR Factorization Part 2 – Algorithm to Find Eigenvalues.
Advertisements

Flows and Networks Plan for today (lecture 2): Questions? Continuous time Markov chain Birth-death process Example: pure birth process Example: pure death.
Markov Chains.
1 Chapter 5 Continuous time Markov Chains Learning objectives : Introduce continuous time Markov Chain Model manufacturing systems using Markov Chain Able.
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
IERG5300 Tutorial 1 Discrete-time Markov Chain
Operations Research: Applications and Algorithms
Use of moment generating functions. Definition Let X denote a random variable with probability density function f(x) if continuous (probability mass function.
Markov Chains 1.
Markov Chain Monte Carlo Prof. David Page transcribed by Matthew G. Lee.
11 - Markov Chains Jim Vallandingham.
TCOM 501: Networking Theory & Fundamentals
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
IEG5300 Tutorial 5 Continuous-time Markov Chain Peter Chen Peng Adapted from Qiwen Wang’s Tutorial Materials.
Андрей Андреевич Марков. Markov Chains Graduate Seminar in Applied Statistics Presented by Matthias Theubert Never look behind you…
Lecture 6 Power spectral density (PSD)
Suggested readings Historical notes Markov chains MCMC details
BAYESIAN INFERENCE Sampling techniques
CS774. Markov Random Field : Theory and Application Lecture 16 Kyomin Jung KAIST Nov
1 CE 530 Molecular Simulation Lecture 8 Markov Processes David A. Kofke Department of Chemical Engineering SUNY Buffalo
Monte Carlo Simulation Methods - ideal gas. Calculating properties by integration.
What if time ran backwards? If X n, 0 ≤ n ≤ N is a Markov chain, what about Y n = X N-n ? If X n follows the stationary distribution, Y n has stationary.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
The Gibbs sampler Suppose f is a function from S d to S. We generate a Markov chain by consecutively drawing from (called the full conditionals). The n’th.
A Bayesian view of language evolution by iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
TCOM 501: Networking Theory & Fundamentals
1 Cluster Monte Carlo Algorithms & softening of first-order transition by disorder TIAN Liang.
If time is continuous we cannot write down the simultaneous distribution of X(t) for all t. Rather, we pick n, t 1,...,t n and write down probabilities.
1 Markov Chains Algorithms in Computational Biology Spring 2006 Slides were edited by Itai Sharon from Dan Geiger and Ydo Wexler.
CS Pattern Recognition Review of Prerequisites in Math and Statistics Prepared by Li Yang Based on Appendix chapters of Pattern Recognition, 4.
Analyzing iterated learning Tom Griffiths Brown University Mike Kalish University of Louisiana.
Group exercise For 0≤t 1
Monte Carlo Methods in Partial Differential Equations.
CS6800 Advanced Theory of Computation Fall 2012 Vinay B Gavirangaswamy
Image Analysis and Markov Random Fields (MRFs) Quanren Xiong.
6. Markov Chain. State Space The state space is the set of values a random variable X can take. E.g.: integer 1 to 6 in a dice experiment, or the locations.
Lecture 11: Ising model Outline: equilibrium theory d = 1
F.F. Assaad. MPI-Stuttgart. Universität-Stuttgart Numerical approaches to the correlated electron problem: Quantum Monte Carlo.  The Monte.
Functions of Random Variables. Methods for determining the distribution of functions of Random Variables 1.Distribution function method 2.Moment generating.
9. Convergence and Monte Carlo Errors. Measuring Convergence to Equilibrium Variation distance where P 1 and P 2 are two probability distributions, A.
Monte Carlo Methods Versatile methods for analyzing the behavior of some activity, plan or process that involves uncertainty.
Borgan and Henderson:. Event History Methodology
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation Radford M. Neal 발표자 : 장 정 호.
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
Markov Chains X(t) is a Markov Process if, for arbitrary times t1 < t2 < < tk < tk+1 If X(t) is discrete-valued If X(t) is continuous-valued i.e.
Chapter 61 Continuous Time Markov Chains Birth and Death Processes,Transition Probability Function, Kolmogorov Equations, Limiting Probabilities, Uniformization.
CS433 Modeling and Simulation Lecture 07 – Part 01 Continuous Markov Chains Dr. Anis Koubâa 14 Dec 2008 Al-Imam.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
CDA6530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
7. Metropolis Algorithm. Markov Chain and Monte Carlo Markov chain theory describes a particularly simple type of stochastic processes. Given a transition.
1/18/2016Atomic Scale Simulation1 Definition of Simulation What is a simulation? –It has an internal state “S” In classical mechanics, the state = positions.
Javier Junquera Importance sampling Monte Carlo. Cambridge University Press, Cambridge, 2002 ISBN Bibliography.
CS774. Markov Random Field : Theory and Application Lecture 15 Kyomin Jung KAIST Oct
6.4 Random Fields on Graphs 6.5 Random Fields Models In “Adaptive Cooperative Systems” Summarized by Ho-Sik Seok.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
COMS Network Theory Week 5: October 6, 2010 Dragomir R. Radev Wednesdays, 6:10-8 PM 325 Pupin Terrace Fall 2010.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Monte Carlo Simulation of Canonical Distribution The idea is to generate states i,j,… by a stochastic process such that the probability  (i) of state.
1 Chapter 5 Continuous time Markov Chains Learning objectives : Introduce continuous time Markov Chain Model manufacturing systems using Markov Chain Able.
Random Sampling Algorithms with Applications Kyomin Jung KAIST Aug ERC Workshop.
How many iterations in the Gibbs sampler? Adrian E. Raftery and Steven Lewis (September, 1991) Duke University Machine Learning Group Presented by Iulian.
The Monte Carlo Method/ Markov Chains/ Metropolitan Algorithm from sec in “Adaptive Cooperative Systems” -summarized by Jinsan Yang.
Advanced Statistical Computing Fall 2016
Industrial Engineering Dep
Monte Carlo Methods in Scientific Computing
Ilan Ben-Bassat Omri Weinstein
Opinionated Lessons #39 MCMC and Gibbs Sampling in Statistics
CS723 - Probability and Stochastic Processes
Presentation transcript:

Lecture 3: Markov processes, master equation Outline: Preliminaries and definitions Chapman-Kolmogorov equation Wiener process Markov chains eigenvectors and eigenvalues detailed balance Monte Carlo master equation

Stochastic processes Random function x(t)

Stochastic processes Random function x(t) Defined by a distribution functional P[x], or by all its moments

Stochastic processes Random function x(t) Defined by a distribution functional P[x], or by all its moments

Stochastic processes Random function x(t) Defined by a distribution functional P[x], or by all its moments or by its characteristic functional:

Stochastic processes Random function x(t) Defined by a distribution functional P[x], or by all its moments or by its characteristic functional:

Stochastic processes (2) Cumulant generating functional:

Stochastic processes (2) Cumulant generating functional:

Stochastic processes (2) Cumulant generating functional: where

Stochastic processes (2) Cumulant generating functional: where correlation function

Stochastic processes (2) Cumulant generating functional: where etc. correlation function

Stochastic processes (3) Gaussian process:

Stochastic processes (3) Gaussian process:

Stochastic processes (3) Gaussian process: (no higher-order cumulants)

Stochastic processes (3) Gaussian process: (no higher-order cumulants) Conditional probabilities:

Stochastic processes (3) Gaussian process: (no higher-order cumulants) Conditional probabilities:

Stochastic processes (3) Gaussian process: (no higher-order cumulants) Conditional probabilities: = probability of x(t1) … x(tk), given x(tk+1) … x(tm)

Wiener-Khinchin theorem Fourier analyze x(t):

Wiener-Khinchin theorem Fourier analyze x(t): Power spectrum:

Wiener-Khinchin theorem Fourier analyze x(t): Power spectrum:

Wiener-Khinchin theorem Fourier analyze x(t): Power spectrum:

Wiener-Khinchin theorem Fourier analyze x(t): Power spectrum:

Wiener-Khinchin theorem Fourier analyze x(t): Power spectrum:

Wiener-Khinchin theorem Fourier analyze x(t): Power spectrum: Power spectrum is Fourier transform of the correlation function

Markov processes No information about the future from past values earlier than the latest available:

Markov processes No information about the future from past values earlier than the latest available:

Markov processes No information about the future from past values earlier than the latest available: Can get general distribution by iterating Q:

Markov processes No information about the future from past values earlier than the latest available: Can get general distribution by iterating Q:

Markov processes No information about the future from past values earlier than the latest available: Can get general distribution by iterating Q: where P(x(t0)) is the initial distribution.

Markov processes No information about the future from past values earlier than the latest available: Can get general distribution by iterating Q: where P(x(t0)) is the initial distribution. Integrate this over x(tn-1), … x(t1) to get

Markov processes No information about the future from past values earlier than the latest available: Can get general distribution by iterating Q: where P(x(t0)) is the initial distribution. Integrate this over x(tn-1), … x(t1) to get

Markov processes No information about the future from past values earlier than the latest available: Can get general distribution by iterating Q: where P(x(t0)) is the initial distribution. Integrate this over x(tn-1), … x(t1) to get The case n = 2 is the

Chapman-Kolmogorov equation

Chapman-Kolmogorov equation

Chapman-Kolmogorov equation (for any t’)

Chapman-Kolmogorov equation (for any t’) Examples: Wiener process (Brownian motion/random walk):

Chapman-Kolmogorov equation (for any t’) Examples: Wiener process (Brownian motion/random walk):

Chapman-Kolmogorov equation (for any t’) Examples: Wiener process (Brownian motion/random walk): (cumulative) Poisson process

Markov chains Both t and x discrete, assuming stationarity

Markov chains Both t and x discrete, assuming stationarity

Markov chains Both t and x discrete, assuming stationarity (because they are probabilities)

Markov chains Both t and x discrete, assuming stationarity (because they are probabilities) Equation of motion:

Markov chains Both t and x discrete, assuming stationarity (because they are probabilities) Equation of motion: Formal solution:

Markov chains (2): properties of T T has a left eigenvector

Markov chains (2): properties of T T has a left eigenvector (because )

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1.

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1. The corresponding right eigenvector is

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1. The corresponding right eigenvector is (the stationary state, because the eigenvalue is 1: )

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1. The corresponding right eigenvector is (the stationary state, because the eigenvalue is 1: ) For all other right eigenvectors with components

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1. The corresponding right eigenvector is (the stationary state, because the eigenvalue is 1: ) For all other right eigenvectors with components

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1. The corresponding right eigenvector is (the stationary state, because the eigenvalue is 1: ) For all other right eigenvectors with components (because they must be orthogonal to : )

Markov chains (2): properties of T T has a left eigenvector (because ) Its eigenvalue is 1. The corresponding right eigenvector is (the stationary state, because the eigenvalue is 1: ) For all other right eigenvectors with components (because they must be orthogonal to : ) All other eigenvalues are < 1.

Detailed balance If there is a stationary distribution P0 with components and

Detailed balance If there is a stationary distribution P0 with components and

Detailed balance If there is a stationary distribution P0 with components and

Detailed balance If there is a stationary distribution P0 with components and can prove (if ergodicity*) convergence to P0 from any initial state:

Detailed balance If there is a stationary distribution P0 with components and can prove (if ergodicity*) convergence to P0 from any initial state: * Can reach any state from any other and no cycles

Detailed balance If there is a stationary distribution P0 with components and can prove (if ergodicity*) convergence to P0 from any initial state: Define , * Can reach any state from any other and no cycles

Detailed balance If there is a stationary distribution P0 with components and can prove (if ergodicity*) convergence to P0 from any initial state: Define , make a similarity transformation * Can reach any state from any other and no cycles

Detailed balance If there is a stationary distribution P0 with components and can prove (if ergodicity*) convergence to P0 from any initial state: Define , make a similarity transformation * Can reach any state from any other and no cycles

Detailed balance If there is a stationary distribution P0 with components and can prove (if ergodicity*) convergence to P0 from any initial state: Define , make a similarity transformation R is symmetric, has complete set of eigenvectors , components (Eigenvalues λj same as those of T.) * Can reach any state from any other and no cycles

Detailed balance (2)

Detailed balance (2)

Detailed balance (2)

Detailed balance (2) Right eigenvectors of T:

Detailed balance (2) Right eigenvectors of T: Now look at evolution:

Detailed balance (2) Right eigenvectors of T: Now look at evolution:

Detailed balance (2) Right eigenvectors of T: Now look at evolution:

Detailed balance (2) Right eigenvectors of T: Now look at evolution:

Detailed balance (2) Right eigenvectors of T: Now look at evolution: (since )

Detailed balance (2) Right eigenvectors of T: Now look at evolution: (since )

Monte Carlo an example of detailed balance

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step,

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random (2) compute “field” of neighbors hi(t) = ΣjJijSj(t)

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random (2) compute “field” of neighbors hi(t) = ΣjJijSj(t) Jij = Jji

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random (2) compute “field” of neighbors hi(t) = ΣjJijSj(t) Jij = Jji (3) Si(t + Δt) = +1 with probability

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random (2) compute “field” of neighbors hi(t) = ΣjJijSj(t) Jij = Jji (3) Si(t + Δt) = +1 with probability

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random (2) compute “field” of neighbors hi(t) = ΣjJijSj(t) Jij = Jji (3) Si(t + Δt) = +1 with probability

Monte Carlo an example of detailed balance Ising model: Binary “spins” Si(t) = ±1 Dynamics: at every time step, (1) choose a spin (i) at random (2) compute “field” of neighbors hi(t) = ΣjJijSj(t) Jij = Jji (3) Si(t + Δt) = +1 with probability (equilibration of Si, given current values of other S’s)

Monte Carlo (2) In language of Markov chains, states (n) are

Monte Carlo (2) In language of Markov chains, states (n) are Single-spin flips: transitions only between neighboring points on hypercube

Monte Carlo (2) In language of Markov chains, states (n) are Single-spin flips: transitions only between neighboring points on hypercube

Monte Carlo (2) In language of Markov chains, states (n) are Single-spin flips: transitions only between neighboring points on hypercube T matrix elements:

Monte Carlo (2) In language of Markov chains, states (n) are Single-spin flips: transitions only between neighboring points on hypercube T matrix elements: all other Tmn = 0.

Monte Carlo (2) In language of Markov chains, states (n) are Single-spin flips: transitions only between neighboring points on hypercube T matrix elements: all other Tmn = 0. Note:

Monte Carlo (2) In language of Markov chains, states (n) are Single-spin flips: transitions only between neighboring points on hypercube T matrix elements: all other Tmn = 0. Note:

Monte Carlo (3) T satisfies detailed balance:

Monte Carlo (3) T satisfies detailed balance: where p0 is the Gibbs distribution:

Monte Carlo (3) T satisfies detailed balance: where p0 is the Gibbs distribution: After many Monte Carlo steps, converge to p0:

Monte Carlo (3) T satisfies detailed balance: where p0 is the Gibbs distribution: After many Monte Carlo steps, converge to p0: S’s sample Gibbs distribution

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm:

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm: If hiSi < 0, Si(t+Δt) = -Si(t),

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm: If hiSi < 0, Si(t+Δt) = -Si(t), If hiSi > 0, Si(t+Δt) = -Si(t) with probability exp(-hiSi)

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm: If hiSi < 0, Si(t+Δt) = -Si(t), If hiSi > 0, Si(t+Δt) = -Si(t) with probability exp(-hiSi) Thus,

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm: If hiSi < 0, Si(t+Δt) = -Si(t), If hiSi > 0, Si(t+Δt) = -Si(t) with probability exp(-hiSi) Thus,

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm: If hiSi < 0, Si(t+Δt) = -Si(t), If hiSi > 0, Si(t+Δt) = -Si(t) with probability exp(-hiSi) Thus, In either case,

Monte Carlo (3): Metropolis version The foregoing was for “heat-bath” MC. Another possibility is the Metropolis algorithm: If hiSi < 0, Si(t+Δt) = -Si(t), If hiSi > 0, Si(t+Δt) = -Si(t) with probability exp(-hiSi) Thus, In either case, i.e., detailed balance with Gibbs

Continuous-time limit: master equation For Markov chain:

Continuous-time limit: master equation For Markov chain:

Continuous-time limit: master equation For Markov chain: Differential equation:

Continuous-time limit: master equation For Markov chain: Differential equation: In components:

Continuous-time limit: master equation For Markov chain: Differential equation: In components: (using normalization of columns of T:)

Continuous-time limit: master equation For Markov chain: Differential equation: In components: (using normalization of columns of T:) (expect , m ≠ n)

Continuous-time limit: master equation For Markov chain: Differential equation: In components: (using normalization of columns of T:) (expect , m ≠ n) transition rate matrix