Untangling equations involving uncertainty Scott Ferson, Applied Biomathematics Vladik Kreinovich, University of Texas at El Paso W. Troy Tucker, Applied.

Slides:



Advertisements
Similar presentations
Fuzzy arithmetic is not a conservative way to do risk analysis Scott Ferson and Lev Ginzburg Applied Biomathematics
Advertisements

Structural reliability analysis with probability- boxes Hao Zhang School of Civil Engineering, University of Sydney, NSW 2006, Australia Michael Beer Institute.
Reliable Dynamic Analysis of Structures Using Imprecise Probability Mehdi Modares and Joshua Bergerson DEPARTMENT OF CIVIL, ARCHITECTURAL AND EVIRONMENTAL.
Comments on Hierarchical models, and the need for Bayes Peter Green, University of Bristol, UK IWSM, Chania, July 2002.
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Psychology 290 Special Topics Study Course: Advanced Meta-analysis April 7, 2014.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Bayesian statistics 2 More on priors plus model choice.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Confidence Intervals for Proportions
Bayesian statistics – MCMC techniques
Planning under Uncertainty
Visual Recognition Tutorial
ECIV 201 Computational Methods for Civil Engineers Richard P. Ray, Ph.D., P.E. Error Analysis.
Pricing an Option Monte Carlo Simulation. We will explore a technique, called Monte Carlo simulation, to numerically derive the price of an option or.
Results 2 (cont’d) c) Long term observational data on the duration of effective response Observational data on n=50 has EVSI = £867 d) Collect data on.
Projecting uncertainty through nonlinear ODEs Youdong Lin 1, Mark Stadherr 1, George Corliss 2, Scott Ferson 3 1 University of Notre Dame 2 Marquette University.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Estimation 8.
- 1 - Summary of P-box Probability bound analysis (PBA) PBA can be implemented by nested Monte Carlo simulation. –Generate CDF for different instances.
Introduction to Bayesian Parameter Estimation
Thanks to Nir Friedman, HU
Maximum likelihood (ML)
Lecture II-2: Probability Review
Standard Error of the Mean
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 31.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Mote Carlo Method for Uncertainty The objective is to introduce a simple (almost trivial) example so that you can Perform.
880.P20 Winter 2006 Richard Kass 1 Confidence Intervals and Upper Limits Confidence intervals (CI) are related to confidence limits (CL). To calculate.
Model Inference and Averaging
Chapter 8: Confidence Intervals
Chapter 7. Statistical Intervals Based on a Single Sample Weiqi Luo ( 骆伟祺 ) School of Software Sun Yat-Sen University : Office.
1 Institute of Engineering Mechanics Leopold-Franzens University Innsbruck, Austria, EU H.J. Pradlwarter and G.I. Schuëller Confidence.
1 Lesson 8: Basic Monte Carlo integration We begin the 2 nd phase of our course: Study of general mathematics of MC We begin the 2 nd phase of our course:
Monte Carlo Methods1 T Special Course In Information Science II Tomas Ukkonen
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
Chapter 1 Overview and Descriptive Statistics 1111.1 - Populations, Samples and Processes 1111.2 - Pictorial and Tabular Methods in Descriptive.
Week 41 Estimation – Posterior mean An alternative estimate to the posterior mode is the posterior mean. It is given by E(θ | s), whenever it exists. This.
Example: Bioassay experiment Problem statement –Observations: At each level of dose, 5 animals are tested, and number of death are observed.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Chapter 3.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 07: BAYESIAN ESTIMATION (Cont.) Objectives:
4 Proposed Research Projects SmartHome – Encouraging patients with mild cognitive disabilities to use digital memory notebook for activities of daily living.
Statistics for Decision Making Basic Inference QM Fall 2003 Instructor: John Seydel, Ph.D.
1 / 12 Michael Beer, Vladik Kreinovich COMPARING INTERVALS AND MOMENTS FOR THE QUANTIFICATION OF COARSE INFORMATION M. Beer University of Liverpool V.
Estimation Method of Moments (MM) Methods of Moment estimation is a general method where equations for estimating parameters are found by equating population.
Machine Learning in Practice Lecture 5 Carolyn Penstein Rosé Language Technologies Institute/ Human-Computer Interaction Institute.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 11 FUZZY.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
- 1 - Computer model under uncertainty In previous lecture on accuracy assessment –We considered mostly deterministic models. – We did not distinguish.
Geology 5670/6670 Inverse Theory 20 Feb 2015 © A.R. Lowry 2015 Read for Mon 23 Feb: Menke Ch 9 ( ) Last time: Nonlinear Inversion Solution appraisal.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Univariate Gaussian Case (Cont.)
CSE 473 Uncertainty. © UW CSE AI Faculty 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one has stood.
Application of Dynamic Programming to Optimal Learning Problems Peter Frazier Warren Powell Savas Dayanik Department of Operations Research and Financial.
Sampling Design and Analysis MTH 494 LECTURE-11 Ossam Chohan Assistant Professor CIIT Abbottabad.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
Outline Historical note about Bayes’ rule Bayesian updating for probability density functions –Salary offer estimate Coin trials example Reading material:
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
NASSP Masters 5003F - Computational Astronomy Lecture 4: mostly about model fitting. The model is our estimate of the parent function. Let’s express.
Univariate Gaussian Case (Cont.)
More about Posterior Distributions
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
LECTURE 09: BAYESIAN LEARNING
LECTURE 07: BAYESIAN ESTIMATION
Distribution-free Monte Carlo for population viability analysis
Untangling equations involving uncertainty
Presentation transcript:

Untangling equations involving uncertainty Scott Ferson, Applied Biomathematics Vladik Kreinovich, University of Texas at El Paso W. Troy Tucker, Applied Biomathematics

Overview Three kinds of operations –Deconvolutions –Backcalculations –Updates (oh, my!) Very elementary methods of interval analysis –Low-dimensional –Simple arithmetic operations But combined with probability theory

Probability box (p-box) Bounds on a cumulative distribution function (CDF) Envelope of a Dempster-Shafer structure Used in risk analysis and uncertainty arithmetic Generalizes probability distributions and intervals This is an interval, not a uniform distribution Cumulative probability

Probability bounds analysis (PBA) assuming independence assuming nothing CDF assuming independence a =T( 0, 10, 20) + [0, 5] b =N([20,23],[1,12]) Disagreement between theoretical and observed variance c = a |+| b c = a + b

PBA handles common problems Imprecisely specified distributions Poorly known or unknown dependencies Non-negligible measurement error Inconsistency in the quality of input data Model uncertainty and non-stationarity Plus, it’s much faster than Monte Carlo

Updating Using knowledge of how variables are related to tighten their estimates Removes internal inconsistency and explicates unrecognized knowledge Also called constraint updating or editing Also called natural extension

Example Suppose W = [23, 33] H = [112, 150] A = [2000, 3200] Does knowing W  H = A let us to say any more?

Answer Yes, we can infer that W = [23, 28.57] H = [112, ] A = [2576, 3200] The formulas are just W = intersect(W, A/H), etc. To get the largest possible W, for instance, let A be as large as possible and H as small as possible, and solve for W =A/H.

Bayesian strategy Prior Likelihood Posterior

Bayes’ rule Concentrates mass onto the manifold of feasible combinations of W, H, and A Answers have the same supports as intervals Computationally complex Needs specification of priors Yields distributions that are not justified (come from the choice of priors) Expresses less uncertainty than is present

Updating with p-boxes A W H

Answers intersect(W, A/H)intersect(H, A/W) intersect(A, W  H)

Calculation with p-boxes Agrees with interval analysis whenever inputs are intervals Relaxes Bayesian strategy when precise priors are not warranted Produces more reasonable answers when priors not well known Much easier to compute than Bayes’ rule

Backcalculation Find constraints on B that ensure C = A + B satisfies specified constraints Or, more generally, C = f(A 1, A 2,…, A k, B) If A and C are intervals, the answer is called the tolerance solution

Can’t just invert the equation When conc is put back into the forward equation, the dose is wider than planned dose  body mass intake conc = conc  intake body mass dose =

Example dose = [0, 2] milligram per kilogram intake = [1, 2.5] liter mass = [60, 96] kilogram conc = dose * mass / intake [ 0, 192] milligram liter -1 dose = conc * intake / mass [ 0, 8] milligram kilogram -1 Doses 4 times larger than tolerable levels!

Backcalculating probability distributions Needed for engineering design problems, e.g., cleanup and remediation planning for environmental contamination Available analytical algorithms are unstable for almost all problems Except in a few special cases, Monte Carlo simulation cannot compute backcalculations; trial and error methods are required

Backcalculation with p-boxes Suppose A + B = C, where A = normal(5, 1) C = {0  C, median  15, 90 th %ile  35, max  50} A

Getting the answer The backcalculation algorithm basically reverses the forward convolution Not hard at all…but a little messy to show Any distribution totally inside B is sure to satisfy the constraint … it’s “kernel” B

Check by plugging back in A + B = C*  C

When you Know that A + B = C A – B = C A  B = C A / B = C A ^ B = C 2A = C A² = C And you have estimates for A, B A, C B,C A, B A, C B,C A, B A, C B,C A, B A, C B,C A, B A, C B,C A C A C Use this formula to find the unknown C = A + B B = backcalc(A,C) A = backcalc (B,C) C = A – B B = –backcalc(A,C) A = backcalc (–B,C) C = A * B B = factor(A,C) A = factor(B,C) C = A / B B = 1/factor(A,C) A = factor(1/B,C) C = A ^ B B = factor(log A, log C) A = exp(factor(B, log C)) C = 2 * A A = C / 2 C = A ^ 2 A = sqrt(C)

Kernels Existence more likely if p-boxes are fat Wider if we can also assume independence Answers are not unique, even though tolerance solutions always are Different kernels can emphasize different properties Envelope of all possible kernels is the shell (i.e., the united solution)

Precise distributions Precise distributions can’t express the nature of the target Finding a conc distribution that results in a prescribed distribution of doses says we want some doses to be high (any distribution to the left would be even better) We need to express the dose target as a p-box

Deconvolution Uses information about dependence to tighten estimates Useful, for instance, in correcting an estimated distribution for measurement uncertainty For instance, suppose Y = X +  If X and  are independent,  Y ² =  X ² +   ² Then we do an uncertainty correction

Example Y = X +  Y,  ~ normal X ~ N(decon(  Y,  X ), sqrt(decon(   ²,  Y ²)) Y ~ N([5,9], [2,3]);  ~ N([  1,+1], [½,1]) X ~ N(dcn([  1,1],[5,6]), sqrt(dcn([¼,1],[4,9]))) X ~ N([6,8], sqrt([3, 63])

Deconvolutions with p-boxes As for backcalculations, computation of deconvolutions is troublesome in probability theory, but often much simpler with p-boxes Deconvolution didn’t have an analog in interval analysis (until now via p-boxes)

Relaxing over-determination Most constraint problems almost never have solutions with probability distributions The constraints are too numerous and strict P-boxes relax these constraints so that many problems can have solutions

P-boxes in interval analysis P-boxes bring probability distributions into the realm of intervals Express and solve backcalculation problems better than is possible in probability theory by itself Generalize the notion of tolerance solutions (kernels) Relax unwarranted assumptions about priors in updating problems needed in a Bayesian approach Introduce deconvolution into interval analysis

Acknowledgments Janos Hajagos, Stony Brook University Lev Ginzburg, Stony Brook University David Myers, Applied Biomathematics National Institutes of Health SBIR program

End

A H A H W