Ordinal Optimization. Copyright by Yu-Chi Ho2 Some Additional Issues F Design dependent estimation noises vs. observation noise F Goal Softening to increase.

Slides:



Advertisements
Similar presentations
Chapter 6 Sampling and Sampling Distributions
Advertisements

CHAPTER 13 M ODELING C ONSIDERATIONS AND S TATISTICAL I NFORMATION “All models are wrong; some are useful.”  George E. P. Box Organization of chapter.
Tuning bandit algorithms in stochastic environments The 18th International Conference on Algorithmic Learning Theory October 3, 2007, Sendai International.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
10/19/2004TCSS435A Isabelle Bichindaritz1 Game and Tree Searching.
Econ 140 Lecture 61 Inference about a Mean Lecture 6.
Chapter 5: Confidence Intervals.
On Systems with Limited Communication PhD Thesis Defense Jian Zou May 6, 2004.
Parameter Estimation: Maximum Likelihood Estimation Chapter 3 (Duda et al.) – Sections CS479/679 Pattern Recognition Dr. George Bebis.
Sampling Distributions
Chapter 7 Sampling and Sampling Distributions
LARGE SAMPLE TESTS ON PROPORTIONS
No Free Lunch (NFL) Theorem Many slides are based on a presentation of Y.C. Ho Presentation by Kristian Nolde.
Part III: Inference Topic 6 Sampling and Sampling Distributions
Chapter 7 Estimation: Single Population
Statistics for Managers Using Microsoft® Excel 7th Edition
The Central Limit Theorem For simple random samples from any population with finite mean and variance, as n becomes increasingly large, the sampling distribution.
Statistics for Managers Using Microsoft® Excel 7th Edition
Slide 1 Tutorial: Optimal Learning in the Laboratory Sciences Working with nonlinear belief models December 10, 2014 Warren B. Powell Kris Reyes Si Chen.
Model Assessment and Selection Florian Markowetz & Rainer Spang Courses in Practical DNA Microarray Analysis.
FDA- A scalable evolutionary algorithm for the optimization of ADFs By Hossein Momeni.
Topic 5 Statistical inference: point and interval estimate
Lecture 14 Dustin Lueker. 2  Inferential statistical methods provide predictions about characteristics of a population, based on information in a sample.
Normal Distributions Z Transformations Central Limit Theorem Standard Normal Distribution Z Distribution Table Confidence Intervals Levels of Significance.
Ran El-Yaniv and Dmitry Pechyony Technion – Israel Institute of Technology, Haifa, Israel Transductive Rademacher Complexity and its Applications.
Performance Measurement n Assignment? n Timing #include double When() { struct timeval tp; gettimeofday(&tp, NULL); return((double)tp.tv_sec + (double)tp.tv_usec.
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 9 Samples.
1 Managing the Computational Cost in a Monte Carlo Simulation by Considering the Value of Information E. Nikolaidis, V. Pandey, Z. P. Mourelatos April.
Utilities and MDP: A Lesson in Multiagent System Based on Jose Vidal’s book Fundamentals of Multiagent Systems Henry Hexmoor SIUC.
Module 14: Confidence Intervals This module explores the development and interpretation of confidence intervals, with a focus on confidence intervals.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
Random Sampling Approximations of E(X), p.m.f, and p.d.f.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Chapter5: Evaluating Hypothesis. 개요 개요 Evaluating the accuracy of hypotheses is fundamental to ML. - to decide whether to use this hypothesis - integral.
Chapter 19 Monte Carlo Valuation. Copyright © 2006 Pearson Addison-Wesley. All rights reserved Monte Carlo Valuation Simulation of future stock.
Framework. Notation r=vector of excess returns –Bold signifies vectors and matrices. We denote individual stock excess returns as r n. –Excess above risk.
Machine Learning Chapter 7. Computational Learning Theory Tom M. Mitchell.
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Chapter 6 Sampling and Sampling Distributions
Review Day 2 May 4 th Probability Events are independent if the outcome of one event does not influence the outcome of any other event Events are.
Confidence Intervals and Sample Size
Copyright © Cengage Learning. All rights reserved.
Sampling Distributions
Statistical Estimation
Visual Recognition Tutorial
12. Principles of Parameter Estimation
Reinforcement Learning (1)
Chapter 7 Sampling Distributions.
Stats Tools for Analyzing Data
Announcements Homework 3 due today (grace period through Friday)
Chapter 7 Sampling Distributions.
Lecture 2 – Monte Carlo method in finance
Computational Learning Theory
Partly Verifiable Signals (c.n.)
Summarizing Data by Statistics
Computational Learning Theory
Chapter 7 Sampling Distributions.
CHAPTER 15 SUMMARY Chapter Specifics
The estimate of the proportion (“p-hat”) based on the sample can be a variety of values, and we don’t expect to get the same value every time, but the.
Chapter 7 Sampling Distributions.
Machine Learning: Lecture 6
Machine Learning: UNIT-3 CHAPTER-1
Evaluating Hypothesis
12. Principles of Parameter Estimation
Chapter 7 Sampling Distributions.
Measuring the Similarity of Rhythmic Patterns
Markov Decision Processes
Markov Decision Processes
Presentation transcript:

Ordinal Optimization

Copyright by Yu-Chi Ho2 Some Additional Issues F Design dependent estimation noises vs. observation noise F Goal Softening to increase the Alignment Probability F Constrained Ordinal Optimization

Copyright by Yu-Chi Ho3 Goal Soften  Performance distribution: Normal(0,  signal )  Additive noise:N(0,  noise ) F # of sample taken:N=100 F The selected set: The best of N samples, i.e., |S|=1 F The good enough set:G=top5%

Copyright by Yu-Chi Ho4 Goal Soften (contd.)

Advanced Ordinal Optimization - Breadth vs. Depth in Searches - Y.C.Ho & X.C. Lin October, 2001 Tsinghua University, China

Copyright by Yu-Chi Ho6 Fundamental Difficulties in the Design of Complex Systems 1. Time consuming performance evaluation via simulation - the 1/(T) 1/2 limit 2. Combinatorially large search space - the NP-hard limit 3. Lack of structure - the No Free Lunch Theorem limit 4. Finite limit of computational resource

Copyright by Yu-Chi Ho7 Premises F It is a finite world! F Computing resources are finite

Copyright by Yu-Chi Ho8 Breadth vs. Depth in Stochastic Optimization F At any point in an optimization process, the question is “How do I spend my next unit of computing resources?” –1. Making sure that the current estimated best is indeed the current best (local or depth) –2. Explore additional points to find a better solution (global or breadth) F Trade off between these two sub-goals to optimize finite resource utilization

Copyright by Yu-Chi Ho9 The “Marriage” Metaphor F You have limited resources in time, money, and energy F You know a given set of possible candidates with imperfect information on each candidate F You also have a preference order on these candidate based on the imperfect information F You can spent the next unit of resources to get better information (=> ordering) about the candidates F Or you can spent the next unit of resources to get more (possibly better) samples, i.e., meet new people. F You can quantify this trade off

Copyright by Yu-Chi Ho10 Definitions & Pb Statement  {  1,...,  M }: M designs sampled   b : best observed design among the M designs   * : true best solution among the M designs  N i : computing resources to be spent on  i F T : computing resources available F G: good enough set  Problem (P) : Max Prob(  b  G) s.t.  N i =T

Copyright by Yu-Chi Ho11 Decomposition of (P)  P(  b  G)  P(  b =  *)P(  *  G)+P(  b  *,  b  G) Probability that the observed best is indeed the true best sampled Probability that the true best sampled is good enough Depth Breadth

Copyright by Yu-Chi Ho12 “Depth” in searches F Each observed performance can be modeled as g.r.v. with mean and variances F variances depends on computing resources  P(J obs (  b ) < J obs (  i )) can be easily calculated as a function of resources allocated to each observation  P(  b =  *)   i  b {P(J obs (  b ) < J obs (  i ))} can be easily lower bounded

Copyright by Yu-Chi Ho13 “Breadth” in searches  Define g=|G|/|  |, e.g., 0.05 => G is top 5 %  P(  *  G) = 1- P(  *  G) = 1-(1-g) M  P(  *  G) as a function of M, samples taken, can be easily evaluated. F More advanced adaptive sampling, e.g., nested partition search

Copyright by Yu-Chi Ho14 Evaluating “g” More Generally  Samples are taken from each partition uniformly but not uniformly over . –Let there be partitions R i i=1,.., k ; U R i =  –g i  |G  R i |/|R i |; M i  samples in R i ; L i =#of observed top designs in R i  P(  *  G) = 1-  (1-g i ) M i  g i(est) =g * (L i / R i )/(L/  ) F can prove that g i(est) is unbiased and consistent

Copyright by Yu-Chi Ho15 Breadth vs. Depth (contd.) F T=computing resources used so far  Depth => Prob (observed best so far= actual best so far)  P(  b =  *)  P d (T)  Breadth => Prob (actual best so far  Good Enough set G)  P(  *  G)  P b (T) F Compare marginal increases to depth and breadth.

Copyright by Yu-Chi Ho16 General and Unified Framework At current state of search with M samples of observed performances J b (  i ), i=1,..., M Calculate r d and r b Decision breadth depth More samples Better estimate of performances

Copyright by Yu-Chi Ho17 Applications F Test problems F Apparel manufacturing problem F Stock Option problem F See X. C. Lin “ A New Approach to Stochastic Optimization” Ph.D.thesis Harvard University, Oct. 2000