Sampling Plans.

Slides:



Advertisements
Similar presentations
Clustering k-mean clustering Genome 559: Introduction to Statistical and Computational Genomics Elhanan Borenstein.
Advertisements

BAYESIAN INFERENCE Sampling techniques
Discrete geometry Lecture 2 1 © Alexander & Michael Bronstein
Searching for the Minimal Bézout Number Lin Zhenjiang, Allen Dept. of CSE, CUHK 3-Oct-2005
Sampling Attila Gyulassy Image Synthesis. Overview Problem Statement Random Number Generators Quasi-Random Number Generation Uniform sampling of Disks,
Recent Development on Elimination Ordering Group 1.
1 Variance Reduction via Lattice Rules By Pierre L’Ecuyer and Christiane Lemieux Presented by Yanzhi Li.
Clustering Color/Intensity
End of Chapter 8 Neil Weisenfeld March 28, 2005.
Maximum Likelihood (ML), Expectation Maximization (EM)
Computer vision: models, learning and inference Chapter 10 Graphical Models.
A) Transformation method (for continuous distributions) U(0,1) : uniform distribution f(x) : arbitrary distribution f(x) dx = U(0,1)(u) du When inverse.
Optimization Methods One-Dimensional Unconstrained Optimization
Monte Carlo Methods in Partial Differential Equations.
Space-Filling DOEs Design of experiments (DOE) for noisy data tend to place points on the boundary of the domain. When the error in the surrogate is due.
Domain decomposition in parallel computing Ashok Srinivasan Florida State University COT 5410 – Spring 2004.
Navigating and Browsing 3D Models in 3DLIB Hesham Anan, Kurt Maly, Mohammad Zubair Computer Science Dept. Old Dominion University, Norfolk, VA, (anan,
1 CE 530 Molecular Simulation Lecture 7 David A. Kofke Department of Chemical Engineering SUNY Buffalo
1 Statistical Mechanics and Multi- Scale Simulation Methods ChBE Prof. C. Heath Turner Lecture 11 Some materials adapted from Prof. Keith E. Gubbins:
Random Number Generators CISC/QCSE 810. What is random? Flip 10 coins: how many do you expect will be heads? Measure 100 people: how are their heights.
1 Lesson 3: Choosing from distributions Theory: LLN and Central Limit Theorem Theory: LLN and Central Limit Theorem Choosing from distributions Choosing.
Design of an Evolutionary Algorithm M&F, ch. 7 why I like this textbook and what I don’t like about it!
Chapter 4 Stochastic Modeling Prof. Lei He Electrical Engineering Department University of California, Los Angeles URL: eda.ee.ucla.edu
The Fundamentals: Algorithms, the Integers & Matrices.
Heuristic Optimization Methods Greedy algorithms, Approximation algorithms, and GRASP.
Quasi-Monte Carlo Methods Fall 2012 By Yaohang Li, Ph.D.
Author: B. C. Bromley Presented by: Shuaiyuan Zhou Quasi-random Number Generators for Parallel Monte Carlo Algorithms.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
A Simulation-Based Study of Overlay Routing Performance CS 268 Course Project Andrey Ermolinskiy, Hovig Bayandorian, Daniel Chen.
1 Microarray Clustering. 2 Outline Microarrays Hierarchical Clustering K-Means Clustering Corrupted Cliques Problem CAST Clustering Algorithm.
01/26/05© 2005 University of Wisconsin Last Time Raytracing and PBRT Structure Radiometric quantities.
Metaheuristics for the New Millennium Bruce L. Golden RH Smith School of Business University of Maryland by Presented at the University of Iowa, March.
8 Sept 2006, DEMA2006Slide 1 An Introduction to Computer Experiments and their Design Problems Tony O’Hagan University of Sheffield.
Lesson 8: Basic Monte Carlo integration
School of Computer Science & Engineering
Heuristic Optimization Methods
SRAM Yield Rate Optimization EE201C Final Project
4. Numerical Integration
Haim Kaplan and Uri Zwick
Path Coupling And Approximate Counting
DSS & Warehousing Systems
Bank-aware Dynamic Cache Partitioning for Multicore Architectures
Clustering (3) Center-based algorithms Fuzzy k-means
Computer Science cpsc322, Lecture 14
Chapter 4a Stochastic Modeling
Multivariate environmental characterization of samples
Computer Science cpsc322, Lecture 14
Maximal Independent Set
5. Quasi-Monte Carlo Method
Efficient Distribution-based Feature Search in Multi-field Datasets Ohio State University (Shen) Problem: How to efficiently search for distribution-based.
Chapter 4: Dynamic Programming
Chapter 4a Stochastic Modeling
Lecture 2 – Monte Carlo method in finance
Chapter 4: Dynamic Programming
Clustering.
Multi-Objective Optimization
Four-Cut: An Approximate Sampling Procedure for Election Audits
≠ Particle-based Variational Inference for Continuous Systems
Monte Carlo Rendering Central theme is sampling:
Lecture 4 - Monte Carlo improvements via variance reduction techniques: antithetic sampling Antithetic variates: for any one path obtained by a gaussian.
Text Categorization Berlin Chen 2003 Reference:
Statistical Methods for Data Analysis Random number generators
VQMC J. Planelles.
Configuration Space of an Articulated Robot
Chapter 4: Dynamic Programming
Clustering.
Random Variables A random variable is a rule that assigns exactly one value to each point in a sample space for an experiment. A random variable can be.
Jaroslav Křivánek, MFF UK
Clustering.
Algorithm Course Algorithms Lecture 3 Sorting Algorithm-1
Presentation transcript:

Sampling Plans

Full Factorial Uniform and evenly spaced samples across domain Simple, easy to implement, and covers domain Sample count grows exponentially with dimension

Random Sampling Uses pseudorandom number generator to define samples according to our desired distribution If variable bounds are known, a common choice is a uniform distribution across domain of possible values Ideally, if enough points are sampled and the right distribution is chosen, the design space will be covered

Uniform Projection Plans Sample locations are random, but their projection onto a particular subspace is uniform

Uniform Projection Plans Examples: Random permutation, Latin square

Stratified Sampling Each point is sampled uniformly at random within each grid cell instead of the center Can capture details that regularly-spaced samples might miss

Space-Filling Metrics A sampling plan may cover a search space fully, but still leave large areas unexplored Space-filling metrics quantify this aspect of sampling-plan performance Example: Uniform Projection

Space-Filling Metrics: Discrepancy The maximum difference between the fraction of samples in a hyper-rectangular subset H and that subset’s volume Often very difficult to compute directly

Space-Filling Metrics: Pairwise Distances Method of measuring relative space-filling performance of two 𝑚-point sampling plans Better spread-out plans will have larger pairwise distances Each set’s pairwise distances are sorted in ascending order. The plan with the first pairwise distance exceeding the other is considered more space-filling Suggests simple algorithm of producing a set of randomly distributed sampling plans, then picking the one with greatest pairwise distances.

Space-Filling Metrics: Morris-Mitchell Alternative to previously suggested algorithm that simplifies optimization problem 𝑑 𝑖 is the 𝑖th pairwise distance and 𝑞>0 is a tunable parameter

Space-Filling Metrics: Morris-Mitchell

Space-Filling Subsets Often, the set of possible sample points is constrained to be a subset of available choices. A space-filling metric for a subset S within a finite set X is the maximum distance between a point in X and the closest point in S using a 𝑝-norm to measure distance. A space-filling subset minimizes this metric. Often computationally intractable, but heuristics like greedy search and exchange often produce acceptable results

Space-Filling Subsets

Quasi-Random Sequences Also called low-discrepancy sequences, quasi-random sequences are deterministic sequences that systematically fill a space such that their integral over the space converges as fast as possible. Used for fast convergence in Monte Carlo integration, which approximates an integral by sampling points in a domain

Quasi-Random Sequences Additive Recurrence: recursively adds irrational numbers Halton Sequence: sequence of fractions generated with coprime numbers Sobol Sequence: recursive XOR operation with carefully chosen numbers

Quasi-Random Sequences

Summary Sampling plans are used to cover search spaces with a limited number of points Full factorial sampling, which involves sampling at the vertices of a uniformly discretized grid, requires a number of points exponential in the number of dimensions Uniform projection plans, which project uniformly over each dimension, can be efficiently generated and can be optimized to be space-filling

Summary Greedy local search and the exchange algorithm can be used to find a subset of points that maximally fill a space Quasi-random sequences are deterministic procedures by which space-filling sampling plans can be generated