Lecture 16 Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals.

Slides:



Advertisements
Similar presentations
Lecture 1 Describing Inverse Problems. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability.
Advertisements

Lecture 20 Continuous Problems Linear Operators and Their Adjoints.
Lecture 10 Nonuniqueness and Localized Averages. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 14 Nonlinear Problems Grid Search and Monte Carlo Methods.
Lecture 13 L1 , L∞ Norm Problems and Linear Programming
Hypothesis testing and confidence intervals by resampling by J. Kárász.
Lecture 23 Exemplary Inverse Problems including Earthquake Location.
Lecture 6 Bootstraps Maximum Likelihood Methods. Boostrapping A way to generate empirical probability distributions Very handy for making estimates of.
Lecture 22 Exemplary Inverse Problems including Filter Design.
Sampling: Final and Initial Sample Size Determination
Lecture 18 Varimax Factors and Empircal Orthogonal Functions.
CHAPTER 8 A NNEALING- T YPE A LGORITHMS Organization of chapter in ISSO –Introduction to simulated annealing –Simulated annealing algorithm Basic algorithm.
Lecture 7 Backus-Gilbert Generalized Inverse and the Trade Off of Resolution and Variance.
Lecture 3 Probability and Measurement Error, Part 2.
Visual Recognition Tutorial
Confidence Intervals Underlying model: Unknown parameter We know how to calculate point estimates E.g. regression analysis But different data would change.
Lecture 2 Probability and Measurement Error, Part 1.
Lecture 5 A Priori Information and Weighted Least Squared.
Lecture 19 Continuous Problems: Backus-Gilbert Theory and Radon’s Problem.
Lecture 15 Nonlinear Problems Newton’s Method. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 4 The L 2 Norm and Simple Least Squares. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Lecture 9 Inexact Theories. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 17 Factor Analysis. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and.
Lecture 21 Continuous Problems Fréchet Derivatives.
Lecture 24 Exemplary Inverse Problems including Vibrational Problems.
Lecture 6 Resolution and Generalized Inverses. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Environmental Data Analysis with MatLab Lecture 5: Linear Models.
G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Statistical Data Analysis: Lecture 2 1Probability, Bayes’ theorem 2Random variables and.
Lecture 3 Review of Linear Algebra Simple least-squares.
Lecture 12 Equality and Inequality Constraints. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Adaptive Rao-Blackwellized Particle Filter and It’s Evaluation for Tracking in Surveillance Xinyu Xu and Baoxin Li, Senior Member, IEEE.
Environmental Data Analysis with MatLab Lecture 3: Probability and Measurement Error.
Lecture 8 The Principle of Maximum Likelihood. Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture.
Environmental Data Analysis with MatLab Lecture 24: Confidence Limits of Spectra; Bootstraps.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
2008 Chingchun 1 Bootstrap Chingchun Huang ( 黃敬群 ) Vision Lab, NCTU.
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Lecture 11 Vector Spaces and Singular Value Decomposition.
Lecture 3: Inferences using Least-Squares. Abstraction Vector of N random variables, x with joint probability density p(x) expectation x and covariance.
Environmental Data Analysis with MatLab Lecture 7: Prior Information.
Lecture II-2: Probability Review
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Simulated Annealing G.Anuradha. What is it? Simulated Annealing is a stochastic optimization method that derives its name from the annealing process used.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
© by Yu Hen Hu 1 ECE533 Digital Image Processing Image Restoration.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
Lab 3b: Distribution of the mean
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Time-dependent Schrodinger Equation Numerical solution of the time-independent equation is straightforward constant energy solutions do not require us.
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Simulated Annealing G.Anuradha.
Molecular Modelling - Lecture 2 Techniques for Conformational Sampling Uses CHARMM force field Written in C++
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Geology 5670/6670 Inverse Theory 20 Feb 2015 © A.R. Lowry 2015 Read for Mon 23 Feb: Menke Ch 9 ( ) Last time: Nonlinear Inversion Solution appraisal.
Geology 5670/6670 Inverse Theory 28 Jan 2015 © A.R. Lowry 2015 Read for Fri 30 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares: Uncertainty The.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
1 Probability and Statistics Confidence Intervals.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
1 Design of experiment for computer simulations Let X = (X 1,…,X p )  R p denote the vector of input values chosen for the computer program Each X j is.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Ch13 Empirical Methods.
Stochastic Methods.
Presentation transcript:

Lecture 16 Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals

Syllabus Lecture 01Describing Inverse Problems Lecture 02Probability and Measurement Error, Part 1 Lecture 03Probability and Measurement Error, Part 2 Lecture 04The L 2 Norm and Simple Least Squares Lecture 05A Priori Information and Weighted Least Squared Lecture 06Resolution and Generalized Inverses Lecture 07Backus-Gilbert Inverse and the Trade Off of Resolution and Variance Lecture 08The Principle of Maximum Likelihood Lecture 09Inexact Theories Lecture 10Nonuniqueness and Localized Averages Lecture 11Vector Spaces and Singular Value Decomposition Lecture 12Equality and Inequality Constraints Lecture 13L 1, L ∞ Norm Problems and Linear Programming Lecture 14Nonlinear Problems: Grid and Monte Carlo Searches Lecture 15Nonlinear Problems: Newton’s Method Lecture 16Nonlinear Problems: Simulated Annealing and Bootstrap Confidence Intervals Lecture 17Factor Analysis Lecture 18Varimax Factors, Empircal Orthogonal Functions Lecture 19Backus-Gilbert Theory for Continuous Problems; Radon’s Problem Lecture 20Linear Operators and Their Adjoints Lecture 21Fréchet Derivatives Lecture 22 Exemplary Inverse Problems, incl. Filter Design Lecture 23 Exemplary Inverse Problems, incl. Earthquake Location Lecture 24 Exemplary Inverse Problems, incl. Vibrational Problems

Purpose of the Lecture Introduce Simulated Annealing Introduce the Bootstrap Method for computing Confidence Intervals

Part 1 Simulated Annealing

Monte Carlo Method completely undirected Newton’s Method completely directed

slow, but foolproof fast, but can fall into local minimum

compromise partially-directed random walk

m (p) E high E medium E low

m (p) E high E medium E low p(m * |m (p) )

m (p) E high E medium E low m*m*

acceptance of m * as m (p+1) always accept in error is smaller accept with probability where T is a parameter if error is bigger

large T 1 always accept m * (undirected random walk) ignores the error completely

small T 0 accept m * only when error is smaller (directed random walk) strictly decreases the error

intermediate T most iterations decrease the error but occasionally allow an m* that increases it

m (p) E high E medium E low large T undirected random walk

m (p) E high E medium E low small T directed random walk

strategy start off with large T slowly decrease T during iterations undirected similar to Monte Carlo method (except more “local”) directed similar to Newton’s method (except precise gradient direction not used)

strategy start off with large T slowly decrease T during iterations claim is that this strategy helps achieve the global minimum more random more directed

analogous to annealing of metals high temperatures atoms randomly moving about due to thermal motions as temperature decreases atoms slowly find themselves in a minimum energy configuration orderly arrangement of a “crystal”

analogous to annealing of metals high temperatures atoms randomly moving about due to thermal motions as temperature decreases atoms slowly find themselves in a minimum energy configuration orderly arrangement of a “crystal” hence “simulated annealing” and T called “temperature”

this is just Metroplois-Hastings (way of producing realizations of a random variable) applied to the p.d.f.

this is just Metroplois-Hastings (way of producing realizations of a random variable) applied to the p.d.f. sampling a distribution that starts out wide and blurry but sharpens up as T is decreases

(A) (B) (C)

for k = [1:Niter] T = 0.1 * Eg0 * ((Niter-k+1)/Niter)^2; ma(1) = random('Normal',mg(1),Dm); ma(2) = random('Normal',mg(2),Dm); da = sin(w0*ma(1)*x) + ma(1)*ma(2); Ea = (dobs-da)'*(dobs-da); if( Ea < Eg ) mg=ma; Eg=Ea; p1his(k+1)=1; else p1 = exp( -(Ea-Eg)/T ); p2 = random('unif',0,1); if( p1 > p2 ) mg=ma; Eg=Ea; end end

Part 2 Bootstrap Method

theory of confidence intervals error is the data result in errors in the estimated model parameters p(d) d d obs p(m) m m est m(d)

theory of confidence intervals error is the data result in errors in the estimated model parameters p(d) d p(m) m 95% confidence m(d) 2½%2½%2½%2½% 2½%2½% 2½%2½% 95% confidence

Gaussian linear theory d = Gm m = G -g d standard error propagation [cov m]=G -g [cov d] G -gT univariate Gaussian distribution has 95% of error within two σ of its mean

What to do with Gaussian nonlinear theory? One possibility linearize theory and use standard error propagation d = g(m) m-m (p) ≈ G (p) –g [d- g(m (p) ) ] [cov m]≈G (p) -g [cov d] G (p) -g

disadvantages unknown accuracy and need to compute gradient of theory G (p) G (p) not computed when using some solution methods

alternative confidence intervals with repeat datasets do the whole experiment many times use results of each experiment to make compute m est create histograms from many m est ’s derive empirical 95% confidence intervals from histograms

Bootstrap Method create approximate repeat datasets by randomly resampling (with duplications) the one existing data set

example of resampling original data set random integers in range 1-6 resampled data set

example of resampling original data set random integers in range 1-6 new data set

example of resampling original data set random integers in range 1-6 resampled data set note repeats

rowindex = unidrnd(N,N,1); xresampled = x( rowindex ); dresampled = dobs( rowindex );

p(d)p’(d) sampling duplication mixing interpretation of resampling

(A) (B) m1m1 m2m2 m1m1 m2m2 p(m 2 ) p(m 1 ) (C)

Nbins=50; m1hmin=min(m1save); m1hmax=max(m1save); Dm1bins = (m1hmax-m1hmin)/(Nbins-1); m1bins=m1hmin+Dm1bins*[0:Nbins-1]'; m1hist = hist(m1save,m1bins); pm1 = m1hist/(Dm1bins*sum(m1hist)); Pm1 = Dm1bins*cumsum(pm1); m1low=m1bins(find(Pm1>0.025,1)); m1high=m1bins(find(Pm1>0.975,1));

Nbins=50; m1hmin=min(m1save); m1hmax=max(m1save); Dm1bins = (m1hmax-m1hmin)/(Nbins-1); m1bins=m1hmin+Dm1bins*[0:Nbins-1]'; m1hist = hist(m1save,m1bins); pm1 = m1hist/(Dm1bins*sum(m1hist)); Pm1 = Dm1bins*cumsum(pm1); m1low=m1bins(find(Pm1>0.025,1)); m1high=m1bins(find(Pm1>0.975,1)); histogram

Nbins=50; m1hmin=min(m1save); m1hmax=max(m1save); Dm1bins = (m1hmax-m1hmin)/(Nbins-1); m1bins=m1hmin+Dm1bins*[0:Nbins-1]'; m1hist = hist(m1save,m1bins); pm1 = m1hist/(Dm1bins*sum(m1hist)); Pm1 = Dm1bins*cumsum(pm1); m1low=m1bins(find(Pm1>0.025,1)); m1high=m1bins(find(Pm1>0.975,1)); empirical p.d.f.

Nbins=50; m1hmin=min(m1save); m1hmax=max(m1save); Dm1bins = (m1hmax-m1hmin)/(Nbins-1); m1bins=m1hmin+Dm1bins*[0:Nbins-1]'; m1hist = hist(m1save,m1bins); pm1 = m1hist/(Dm1bins*sum(m1hist)); Pm1 = Dm1bins*cumsum(pm1); m1low=m1bins(find(Pm1>0.025,1)); m1high=m1bins(find(Pm1>0.975,1)); empirical c.d.f.

Nbins=50; m1hmin=min(m1save); m1hmax=max(m1save); Dm1bins = (m1hmax-m1hmin)/(Nbins-1); m1bins=m1hmin+Dm1bins*[0:Nbins-1]'; m1hist = hist(m1save,m1bins); pm1 = m1hist/(Dm1bins*sum(m1hist)); Pm1 = Dm1bins*cumsum(pm1); m1low=m1bins(find(Pm1>0.025,1)); m1high=m1bins(find(Pm1>0.975,1)); 95% confidence bounds