Bayesian Travel Time Reliability

Slides:



Advertisements
Similar presentations
Introduction to Monte Carlo Markov chain (MCMC) methods
Advertisements

Bayesian Estimation in MARK
1 Parametric Sensitivity Analysis For Cancer Survival Models Using Large- Sample Normal Approximations To The Bayesian Posterior Distribution Gordon B.
Markov-Chain Monte Carlo
1 Vertically Integrated Seismic Analysis Stuart Russell Computer Science Division, UC Berkeley Nimar Arora, Erik Sudderth, Nick Hay.
Yanxin Shi 1, Fan Guo 1, Wei Wu 2, Eric P. Xing 1 GIMscan: A New Statistical Method for Analyzing Whole-Genome Array CGH Data RECOMB 2007 Presentation.
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
Bayesian Analysis of X-ray Luminosity Functions A. Ptak (JHU) Abstract Often only a relatively small number of sources of a given class are detected in.
The University of Texas at Austin, CS 395T, Spring 2008, Prof. William H. Press IMPRS Summer School 2009, Prof. William H. Press 1 4th IMPRS Astronomy.
Nonlinear and Non-Gaussian Estimation with A Focus on Particle Filters Prasanth Jeevan Mary Knox May 12, 2006.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Bayesian Analysis for Extreme Events Pao-Shin Chu and Xin Zhao Department of Meteorology School of Ocean & Earth Science & Technology University of Hawaii-
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Standard error of estimate & Confidence interval.
Bayes Factor Based on Han and Carlin (2001, JASA).
Overview G. Jogesh Babu. Probability theory Probability is all about flip of a coin Conditional probability & Bayes theorem (Bayesian analysis) Expectation,
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
Confidence Intervals (Chapter 8) Confidence Intervals for numerical data: –Standard deviation known –Standard deviation unknown Confidence Intervals for.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Confidence Interval Estimation.
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
Machine Learning Lecture 23: Statistical Estimation with Sampling Iain Murray’s MLSS lecture on videolectures.net:
Finding Scientific topics August , Topic Modeling 1.A document as a probabilistic mixture of topics. 2.A topic as a probability distribution.
1 Gil McVean Tuesday 24 th February 2009 Markov Chain Monte Carlo.
Likelihood and Bayes Brian O’Meara
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
A Comparison of Two MCMC Algorithms for Hierarchical Mixture Models Russell Almond Florida State University College of Education Educational Psychology.
Center for Radiative Shock Hydrodynamics Fall 2011 Review Assessment of predictive capability Derek Bingham 1.
An Efficient Sequential Design for Sensitivity Experiments Yubin Tian School of Science, Beijing Institute of Technology.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
1 OUTPUT ANALYSIS FOR SIMULATIONS. 2 Introduction Analysis of One System Terminating vs. Steady-State Simulations Analysis of Terminating Simulations.
Learning to Detect Events with Markov-Modulated Poisson Processes Ihler, Hutchins and Smyth (2007)
1 A Comparison of Information Management using Imprecise Probabilities and Precise Bayesian Updating of Reliability Estimates Jason Matthew Aughenbaugh,
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
MCMC reconstruction of the 2 HE cascade events Dmitry Chirkin, UW Madison.
- 1 - Matlab statistics fundamentals Normal distribution % basic functions mew=100; sig=10; x=90; normpdf(x,mew,sig) 1/sig/sqrt(2*pi)*exp(-(x-mew)^2/sig^2/2)
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
CSC321: Introduction to Neural Networks and Machine Learning Lecture 17: Boltzmann Machines as Probabilistic Models Geoffrey Hinton.
Bayesian II Spring Major Issues in Phylogenetic BI Have we reached convergence? If so, do we have a large enough sample of the posterior?
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
LESSON 5 - STATISTICS & RESEARCH STATISTICS – USE OF MATH TO ORGANIZE, SUMMARIZE, AND INTERPRET DATA.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
A Study on Speaker Adaptation of Continuous Density HMM Parameters By Chin-Hui Lee, Chih-Heng Lin, and Biing-Hwang Juang Presented by: 陳亮宇 1990 ICASSP/IEEE.
Canadian Bioinformatics Workshops
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Reducing Photometric Redshift Uncertainties Through Galaxy Clustering
Bayesian data analysis
Variational Bayes Model Selection for Mixture Distribution
Classification of unlabeled data:
Michael Epstein, Ben Calderhead, Mark A. Girolami, Lucia G. Sivilotti 
Multidimensional Integration Part I
Confidence Intervals Chapter 10 Section 1.
Ch13 Empirical Methods.
Additional notes on random variables
Additional notes on random variables
Ensemble forecasts and seasonal precipitation tercile probabilities
CS639: Data Management for Data Science
Presentation transcript:

Bayesian Travel Time Reliability Feng Guo, Associate Professor Department of Statistics, Virginia Tech Virginia Tech Transportation Institute Dengfeng Zhang, Department of Statistics, Virginia Tech

Travel Time Reliability Travel time is random in nature. Effects to quantify the uncertainty Percentage Variation Misery index Buffer time index Distribution Normal Log-normal distribution …

Multi-State Travel Time Reliability Models Better fitting for the data Easy for interpretation and prediction, similar to weather forecasting: The probability of encountering congestion The estimated travel time IF congestion Guo et al 2010

Multi-State Travel Time Reliability Models Direct link with underline traffic condition and fundamental diagram Can be extended to skewed component distributions such as log-normal Park et al 2010; Guo et al 2012

Model Specification 𝑓 𝑦 ~𝜆 𝑓 1 𝑦 𝜇 1 , 𝜎 1 + 1−𝜆 𝑓 2 (𝑦| 𝜇 2 , 𝜎 2 ) 𝑓 𝑦 ~𝜆 𝑓 1 𝑦 𝜇 1 , 𝜎 1 + 1−𝜆 𝑓 2 (𝑦| 𝜇 2 , 𝜎 2 ) 𝑓 1 : distribution under free-flow condition 𝑓 2 : distribution under congested condition 𝜆 is the proportion of the free-flow component.

Model Parameters Vary by Time of Day Mean Variance 90th Percentile Probability in congested state What is the root cause of this fluctuation?

Bayesian Multi-State Travel Time Models The fluctuation by time-of-day most like due to traffic volume How to incorporate this into the model?

Model Specification: Model 1 𝑓 𝑦 ~𝜆 𝑓 1 𝑦 𝜇 1 , 𝜎 1 + 1−𝜆 𝑓 2 (𝑦| 𝜇 2 , 𝜎 2 ) 𝑓 1 : distribution under free-flow condition 𝑓 2 : distribution under congested condition 𝜆 is the proportion of the free-flow component. 𝜇 2 = 𝜃 0 + 𝜃 1 ∗𝑥 Φ −1 1−𝜆 = 𝛽 0 + 𝛽 1 ∗𝑥 Link mean travel time of congested state with covariates Link probability of travel time state with covariates

Bayesian Model Setup Inference based on posterior distribution Using non-informative priors: let data dominate results. Developed Markov China Monte Carlo (MCMC) to simulate posterior distributions

Issues with Model 1 When traffic volume is low, the two component distribution can be very similar to each other The mixture proportion estimation is not stable

Model Specification: Model 2 𝑓 𝑦 ~𝜆 𝑓 1 𝑦 𝜇 1 , 𝜎 1 + 1−𝜆 𝑓 2 (𝑦| 𝜇 2 , 𝜎 2 ) 𝑓 1 : distribution under free-flow condition 𝑓 2 : distribution under congested condition 𝜆 is the proportion of the free-flow component. 𝜇 2 = 𝜽 𝒔 ∗ 𝝁 𝟏 + 𝜃 1 ∗𝑥 Φ −1 1−𝜆 = 𝛽 0 + 𝛽 1 ∗𝑥 where 𝜽 𝒔 is a predefined scale parameter: How large the minimum value of 𝝁 𝟐 comparing to 𝝁 𝟏

Comparing Model 1 and 2 𝜇 2 = 𝜽 𝒔 ∗ 𝝁 𝟏 + 𝜃 1 ∗𝑥 𝜽 𝒔 =1: the minimum value of congested state is the same as free flow 𝜽 𝒔 =1.5: congested state is at least 50% higher than free flow 𝜽 𝒔 -1

Simulation Study To evaluate the performance of models Based on two metrics Average of posterior mean Coverage probability Set n=Number of simulations we plan to run. For (i in 1:n) { Generate data Do Markov Chain Monte Carlo }While convergence Record if the 95% credible intervals cover the true values }

Simulation Study: Data Generation 𝑋 𝑖𝑗 : Observed traffic volume at time interval i on day j 𝜇 𝑖 : Average Traffic volume at time interval i (e.g. 8:00-9:00) 𝜇 𝑖 = 𝑘 𝑋 𝑖𝑘 𝑁𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑑𝑎𝑦𝑠 𝑌 𝑖𝑗 : Simulated traffic volume at time interval i on day j 𝑌 𝑖𝑗 = 𝑐∗ 𝜇 𝑖 + 𝜖 𝑖𝑗 + , 𝜖 𝑖𝑗 ~𝑁(0, 𝑑 2 )

Model 1 VS Model 2: Posterior Means

Model 1 VS Model 2: Coverage Probability Setting 3: when both 𝜃 0 and 𝜃 1 are small

Robustness What if… the true value of 𝜃 𝑠 is unknown The two components are too close We showed that the overall estimation are quite stable, even if the tuning parameter is misspecified When the two components are too close, by selecting a misspecified tuning parameter could improve the coverage probabilities of some parameters

Robustness

Robustness

Robustness: Coverage Probabilities

Data The data set contains 4306 observations A section of the I-35 freeway in San Antonio, Texas. Vehicles were tagged by a radio frequency device High precision

Study Corridor

Modeling Results

Real Data Analysis

Probability of Congested State as A Function of Traffic Volume

Next Step… Apply the model to a large dataset Any available data are welcome! Hidden Markov Model

HMM The models discussed are based on the assumption that all the observations are independent. Is it realistic?

Hidden Markov Model Hidden Markov model is able to incorporate the dependency structure of the data. Markov chain is a sequence satisfies: 𝑃 𝑥 𝑡+1 =𝑥 𝑥 1 , 𝑥 2 ,… 𝑥 𝑡 =𝑃 𝑥 𝑡+1 =𝑥 𝑥 𝑡 In hidden Markov Chain, the state 𝑥 𝑡 is not visible (i.e. latent) but the output 𝑦 𝑡 is determined by 𝑥 𝑡

Hidden Markov Model Latent state: 𝑤 𝑡 = 1:𝑓𝑟𝑒𝑒−𝑓𝑙𝑜𝑤 2:𝑐𝑜𝑛𝑔𝑒𝑠𝑡𝑒𝑑 𝑤 𝑡 = 1:𝑓𝑟𝑒𝑒−𝑓𝑙𝑜𝑤 2:𝑐𝑜𝑛𝑔𝑒𝑠𝑡𝑒𝑑 Distribution of travel time: 𝑦 𝑡 | 𝑤 𝑡 ~ 𝑓 1 , 𝑖𝑓 𝑤 𝑡 =1 𝑓 2 , 𝑖𝑓 𝑤 𝑡 =2 𝑤 𝑡 and 𝑤 𝑡−1 satisfy Markov property: 𝑃 𝑤 𝑡+1 =𝑤 𝑤 1 , 𝑤 2 ,… 𝑤 𝑡 =𝑃 𝑤 𝑡+1 =𝑤 𝑤 𝑡 If { 𝑤 𝑡 } are independent, this is exactly the traditional mixture Gaussian model we have discussed.

Model Specification Transition Probability: 𝑃 𝑖𝑗 =𝑃( 𝑤 𝑡+1 =𝑗| 𝑤 𝑡 =𝑖) E.g. 𝑃 12 is the probability that the travel time is jumping from free-flow state to congested state. We use logit link function to model the transition probabilities with traffic volume: log 𝑃 12 𝑃 11 = 𝛽 0,1 + 𝛽 1,1 ∗𝑥 log 𝑃 22 𝑃 21 = 𝛽 0,2 + 𝛽 1,2 ∗𝑥

Preliminary Results When the traffic volume is higher, the congested state will be more likely to stay and free-flow state will be more likely to make a jump. The mean travel time of the two states are 578.8 and 972.6 seconds. If we calculate the stationary distribution, the proportion of congested state is around 11.3%. AIC indicates that hidden Markov model is superior to traditional mixture Gaussian model.

Simulation Study

Questions? … Thanks!