Optimization under Uncertainty

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

General Linear Model With correlated error terms  =  2 V ≠  2 I.
Treatments of Risks and Uncertainty in Projects The availability of partial or imperfect information about a problem leads to two new category of decision-making.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Sampling: Final and Initial Sample Size Determination
Uncertainty in Engineering The presence of uncertainty in engineering is unavoidable. Incomplete or insufficient data Design must rely on predictions or.
Integration of sensory modalities
All Hands Meeting, 2006 Title: Grid Workflow Scheduling in WOSE (Workflow Optimisation Services for e- Science Applications) Authors: Yash Patel, Andrew.
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
Visual Recognition Tutorial
Classification and risk prediction
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 3, 4, 6, 7.
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Statistical Decision Theory, Bayes Classifier
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 4, 6, 7.
1 Sampling Models for the Population Mean Ed Stanek UMASS Amherst.
Visual Recognition Tutorial
1 Wavelet synopses with Error Guarantees Minos Garofalakis Phillip B. Gibbons Information Sciences Research Center Bell Labs, Lucent Technologies Murray.
Jeff Howbert Introduction to Machine Learning Winter Classification Bayesian Classifiers.
Today Wrap up of probability Vectors, Matrices. Calculus
CHAPTER 15 S IMULATION - B ASED O PTIMIZATION II : S TOCHASTIC G RADIENT AND S AMPLE P ATH M ETHODS Organization of chapter in ISSO –Introduction to gradient.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
01/24/05© 2005 University of Wisconsin Last Time Raytracing and PBRT Structure Radiometric quantities.
EM and expected complete log-likelihood Mixture of Experts
Statistical Decision Theory
Geo479/579: Geostatistics Ch12. Ordinary Kriging (1)
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 02: BAYESIAN DECISION THEORY Objectives: Bayes.
Approximation Algorithms Pages ADVANCED TOPICS IN COMPLEXITY THEORY.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Statistical Methods II&III: Confidence Intervals ChE 477 (UO Lab) Lecture 5 Larry Baxter, William Hecker, & Ron Terry Brigham Young University.
Statistical Methods II: Confidence Intervals ChE 477 (UO Lab) Lecture 4 Larry Baxter, William Hecker, & Ron Terry Brigham Young University.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Ch15: Decision Theory & Bayesian Inference 15.1: INTRO: We are back to some theoretical statistics: 1.Decision Theory –Make decisions in the presence of.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
MANAGEMENT SCIENCE The Art of Modeling with Spreadsheets STEPHEN G. POWELL KENNETH R. BAKER Compatible with Analytic Solver Platform FOURTH EDITION OPTIMIZATION.
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
1 BA 555 Practical Business Analysis Linear Programming (LP) Sensitivity Analysis Simulation Agenda.
1 Optimizing Decisions over the Long-term in the Presence of Uncertain Response Edward Kambour.
Machine Learning 5. Parametric Methods.
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
Geology 5670/6670 Inverse Theory 27 Feb 2015 © A.R. Lowry 2015 Read for Wed 25 Feb: Menke Ch 9 ( ) Last time: The Sensitivity Matrix (Revisited)
A Brief Maximum Entropy Tutorial Presenter: Davidson Date: 2009/02/04 Original Author: Adam Berger, 1996/07/05
Chance Constrained Robust Energy Efficiency in Cognitive Radio Networks with Channel Uncertainty Yongjun Xu and Xiaohui Zhao College of Communication Engineering,
1 Information Content Tristan L’Ecuyer. 2 Degrees of Freedom Using the expression for the state vector that minimizes the cost function it is relatively.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Kriging - Introduction Method invented in the 1950s by South African geologist Daniel Krige (1919-) for predicting distribution of minerals. Became very.
Objectives: Loss Functions Risk Min. Error Rate Class. Resources: DHS – Chap. 2 (Part 1) DHS – Chap. 2 (Part 2) RGO - Intro to PR MCE for Speech MCE for.
Lecture 1.31 Criteria for optimal reception of radio signals.
Chapter 7. Classification and Prediction
Probability Theory and Parameter Estimation I
LECTURE 03: DECISION SURFACES
Chapter 17: Optimization In Simulation
Robust Fisher Discriminant Analysis
Introduction to Instrumentation Engineering
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
Integration of sensory modalities
Biointelligence Laboratory, Seoul National University
Image and Video Processing
Parametric Methods Berlin Chen, 2005 References:
Mathematical Foundations of BME Reza Shadmehr
Information Theoretical Analysis of Digital Watermarking
Dr. Arslan Ornek MATHEMATICAL MODELS
Probabilistic Surrogate Models
Uncertainty Propagation
Presentation transcript:

Optimization under Uncertainty

Uncertainty Irreducible uncertainty is inherent to a system Epistemic uncertainty is caused by the subjective lack of knowledge by the algorithm designer In optimization problems, uncertainty can be represented by a vector of random variables z over which the designer has no control Feasibility now depends on (x , z) Uncertainty can be part of the design points, objective function, and/or constraints

Uncertainty Design choices are affected by uncertainty

Set-Based Uncertainty Set-based uncertainty treats z as belonging to a set Z Often used to solve problems with minimax approach The minimax approach seeks to minimize f(x,z) assuming a worst-case value of z

Set-Based Uncertainty: Example Consider objective function where 𝑥 =𝑥+𝑧, with set-based uncertainty region 𝑧∈[−𝜖,𝜖] The minimax approach is a minimization problem over the modified objective function f 𝑚𝑜𝑑 𝑥 = maximize 𝑧∈ −𝜖,𝜖 f(𝑥,𝑧)

Set-Based Uncertainty: Example

Set-Based Uncertainty Information-gap decision theory parameterizes the uncertainty set by a nonnegative scalar gap parameter ϵ This gap parameter controls the volume of the uncertainty set Z(ϵ) The goal in information-gap decision theory is to find the design point that is both feasible and allows the largest possible uncertainty gap Searches for the most robust designs

Set-Based Uncertainty: Example Consider robust optimization of f 𝑥,𝑧 = 𝑥 2 +5 𝑒 − 𝑥 2 with 𝑥 = x + z subject to the constraint 𝑥 ∈[ −2 ,2 ] with the uncertainty set 𝒵 𝜖 =[−𝜖,𝜖]

Probabilistic Uncertainty Probabilistic uncertainty uses distributions over the set Z Based on expert knowledge or learned from data Given a distribution p over Z, this section covers five different metrics that convert distributions to scalar values to be optimized

Probabilistic Uncertainty: Expected Value A simple scalar representation of a distribution is by the expected value or mean Often this integral is not computable analytically, but other techniques will be presented in Ch 18

Probabilistic Uncertainty: Expected Value A common model is to apply zero-mean Gaussian noise to the function output, f (x,z) = f (x) + z The expected value is equivalent to the noise-free case

Probabilistic Uncertainty: Expected Value Noise can also be added to the design vector, f (x,z) = f (x+z) = f ( x ) In this case, the expected value is affected by the variance of the zero-mean Gaussian noise

Probabilistic Uncertainty: Expected Value Consider minimizing the expected value of f( x ) = sin(2 x )/ x with 𝑥 =𝑥+𝑧 where z is drawn from zero-mean Gaussian distribution with variance ν.

Probabilistic Uncertainty: Variance When minimized, corresponds to design points that are not overly sensitive to uncertainty Design points with large variance are called sensitive and those with small variance are robust

Probabilistic Uncertainty Consider the function f(x,z) = x2+z with z drawn from a Gamma distribution that depends on x. This distribution has mean 4/(1+|x|) and variance 8/(1+|x|)

Probabilistic Uncertainty

Probabilistic Uncertainty: Statistical Feasibility The probability that a design point is feasible Can be estimated through sampling This metric is maximized

Probabilistic Uncertainty: Value at Risk The best objective value that can be guaranteed with probability α Expressed in terms of the cumulative distribution formula, Φ(y) which defines the probability that the outcome is less than or equal to y VaR is the minimum value of y such that Φ(y) ≥ α Conditional value at risk is the expected value of the top 1 - α quantile of a probability distribution over the output

Probabilistic Uncertainty: Value at Risk

Summary Uncertainty in the optimization process can arise due to errors in the data, the models, or the optimization method itself Accounting for these sources of uncertainty is important in ensuring robust designs Optimization with respect to set-based uncertainty includes the minimax approach that assumes the worst-case and information-gap decision theory that finds a design robust to a maximally sized uncertainty set Probabilistic approaches typically minimize the expected value, the variance, risk of infeasibility, value at risk, conditional value at risk, or a combination of these