Stochastic Monte Carlo methods for non-linear statistical inverse problems Benjamin R. Herman Department of Electrical Engineering City College of New.

Slides:



Advertisements
Similar presentations
EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
Advertisements

ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Bayesian Estimation in MARK
Probability and Statistics Mat 271E
Ch 11. Sampling Models Pattern Recognition and Machine Learning, C. M. Bishop, Summarized by I.-H. Lee Biointelligence Laboratory, Seoul National.
Markov-Chain Monte Carlo
Computer Vision Lab. SNU Young Ki Baik An Introduction to MCMC for Machine Learning (Markov Chain Monte Carlo)
STAT 497 APPLIED TIME SERIES ANALYSIS
Markov Chains 1.
11 - Markov Chains Jim Vallandingham.
Introduction to Sampling based inference and MCMC Ata Kaban School of Computer Science The University of Birmingham.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Lecture 3: Markov processes, master equation
Bayesian Reasoning: Markov Chain Monte Carlo
Bayesian statistics – MCMC techniques
Suggested readings Historical notes Markov chains MCMC details
BAYESIAN INFERENCE Sampling techniques
1 CE 530 Molecular Simulation Lecture 8 Markov Processes David A. Kofke Department of Chemical Engineering SUNY Buffalo
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
Machine Learning CUNY Graduate Center Lecture 7b: Sampling.
Sérgio Pequito Phd Student
Today Introduction to MCMC Particle filters and MCMC
Lecture II-2: Probability Review
Review of Probability.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Bayes Factor Based on Han and Carlin (2001, JASA).
Geneva, September 2010 EARLINET-ASOS Symposium Second GALION Workshop Uncertainties evaluation for aerosol optical properties Aldo Amodeo CNR-IMAA.
Material Model Parameter Identification via Markov Chain Monte Carlo Christian Knipprath 1 Alexandros A. Skordos – ACCIS,
01/24/05© 2005 University of Wisconsin Last Time Raytracing and PBRT Structure Radiometric quantities.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Nonlinear Data Assimilation and Particle Filters
Chapter 14 Monte Carlo Simulation Introduction Find several parameters Parameter follow the specific probability distribution Generate parameter.
Computing a posteriori covariance in variational DA I.Gejadze, F.-X. Le Dimet, V.Shutyaev.
SIS Sequential Importance Sampling Advanced Methods In Simulation Winter 2009 Presented by: Chen Bukay, Ella Pemov, Amit Dvash.
Elements of Financial Risk Management Second Edition © 2012 by Peter Christoffersen 1 Distributions and Copulas for Integrated Risk Management Elements.
CSDA Conference, Limassol, 2005 University of Medicine and Pharmacy “Gr. T. Popa” Iasi Department of Mathematics and Informatics Gabriel Dimitriu University.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
Forward-Scan Sonar Tomographic Reconstruction PHD Filter Multiple Target Tracking Bayesian Multiple Target Tracking in Forward Scan Sonar.
R I T Rochester Institute of Technology Photon Mapping Development LEO-15 DATA SETS (AVIRIS, IOPs) Atmospheric Compensation HYDROLIGHT Simulations (LUT)
Fig. 3 Wind measurements experimental setup Lidar (light detection and ranging) operates using the same concept of microwave RADAR, but it employs a lot.
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation Radford M. Neal 발표자 : 장 정 호.
Clustering and Testing in High- Dimensional Data M. Radavičius, G. Jakimauskas, J. Sušinskas (Institute of Mathematics and Informatics, Vilnius, Lithuania)
1 Sample Geometry and Random Sampling Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Maximum a posteriori sequence estimation using Monte Carlo particle filters S. J. Godsill, A. Doucet, and M. West Annals of the Institute of Statistical.
Improved Cross Entropy Method For Estimation Presented by: Alex & Yanna.
Geology 6600/7600 Signal Analysis 02 Sep 2015 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
Application of the MCMC Method for the Calibration of DSMC Parameters James S. Strand and David B. Goldstein The University of Texas at Austin Sponsored.
STOCHASTIC HYDROLOGY Stochastic Simulation of Bivariate Distributions Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National.
Point Distribution Models Active Appearance Models Compilation based on: Dhruv Batra ECE CMU Tim Cootes Machester.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Stochastic Hydrology Random Field Simulation Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
Review of statistical modeling and probability theory Alan Moses ML4bio.
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Kalman Filter with Process Noise Gauss- Markov.
Gil McVean, Department of Statistics Thursday February 12 th 2009 Monte Carlo simulation.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Canadian Bioinformatics Workshops
Optimization of Monte Carlo Integration
Basic simulation methodology
Markov chain monte carlo
EE513 Audio Signals and Systems
Yalchin Efendiev Texas A&M University
Uncertainty Propagation
Presentation transcript:

Stochastic Monte Carlo methods for non-linear statistical inverse problems Benjamin R. Herman Department of Electrical Engineering City College of New York

Overview Aerosol lidar system Retrieval of aerosol properties from lidar measurements Metropolis-Hastings algorithm for uncertainty analysis Techniques for its efficient implementation

CCNY Aerosol/Water vapor lidar Transmits light at 1064, 532, and 355nm Detects signals at transmitted wavelengths as well as two wavelengths shifted from 355nm by Raman scattering from N2 and H2O 1: Three wavelength pulsed Nd:YAG Laser 2: Newtonian Telescope 3: Receiver optics and photo-detectors 4: Licel transient recorder, analog and photon counting

Optical coefficients and lidar equations Elastic lidar equation Raman lidar equation

λ=1064nm λ=355nm λ=532nm Multi-mode aerosol model and optical coefficients Aerosol is modeled as being composed of different constituents, each with a size distribution and index of refraction Volume concentration Median Radius Geometric standard deviation Refractive index Optical coefficient dependency Aerosol state vector

How can uncertainty in the high dimension aerosol model be understood to assess uncertainty of aerosol properties? Uncertainty analysis Estimated optical coefficients (retrieval results from lidar signal processing) Estimation error Uncertainty is expressed in the form of a posterior probability density function (PDF)

Stochastic Monte Carlo Sequence of Random Variables Θ n with PDFs f n Markov Chain Stationary distributions Monte Carlo simulation Begin with Θ 0 and generate subsequent Θ according to some K

Metropolis-Hastings Algorithm Designed so that a target distribution π(θ) is stationary 1) Generate candidate Φ given current outcome Θ from a random number generator congruent with a proposal distribution whose PDF is q(φ|θ) 2) Accept Φ as the next outcome with probability, 3) If Φ is not accepted then Θ is retained as the next outcome Initial distributions f 0 (θ) will converge to π(θ) with only lose requirements of the proposal distribution. Its choice affects the rate of convergence but generally not the stationary distribution. Types of Proposal Distributions Random walk generators Independent generators

Implementation Ensemble of parallel chains starting from an initial distribution Random walk generation for first several links Ensemble is analyzed partway along the chain, clusters are determined and used to form a hybrid proposal distribution Hybrid proposal distribution Randomly choose walk generation or generation from a distribution corresponding to one of the clusters, thereby allowing teleportation between regions. This aspect is essential for dealing with multiple local maxima in the posterior PDF.

Ensemble evolution animation Example in a reduced single mode aerosol size distribution The target distribution PDF is only a function of median radius and geometric standard deviation

Speeding convergence by initial resampling Ensemble members are not getting represented by clusters Results in tendency not to teleport Hybrid B also includes a small probability of sampling from the initial distribution

Comparison with forward Monte Carlo analysis Numerically integrated posterior cumulative distribution functions (CDFs) (Black) and CDFs estimated from the simple forward Monte Carlo method applied using the maximum a posteriori inverse (Blue) and the Ensemble Metropolis-Hastings algorithm (Red).

Extension to bi-modal particle size distributions Speeding convergence by fitting proposal PDF to the target PDF Eigenvalue conditioning S must be positive definite to be a valid covariance matrix. This is not guaranteed if the target distribution is not Gaussian, and so it is remedied through eigenvector decomposition. Negative Eigenvalues are set to an upper limit and S is recomposed. First derivative fitting Second derivative fitting Multivariate Gaussian PDF

Implementations Higher acceptance probability with uncorrelated random walk is due to scaling down the fitted variance Acceptance rate is used for determining at which link to cluster the ensemble to form the hybrid proposal distribution Full second order derivative fitting compared with uncorrelated partially fitted random walk generation

After links there still remain some members around a local maximum Link 1000 All ensemble members had been together since ~link 500 Hybridization only once Periodic re-hybridization to speed convergence

Things to think about How to choose when to form the multi- component independent generator from clusters Can the independent generator be periodically reformulated indefinitely while still maintaining convergence to the target distribution Which clustering method works best Is there an optimal compromise between random walk with uncorrelated variance versus complete fitting of all mixed derivatives