The Poincaré Constant of a Random Walk in High- Dimensional Convex Bodies Ivona Bezáková Thesis Advisor: Prof. Eric Vigoda.

Slides:



Advertisements
Similar presentations
Sublinear-time Algorithms for Machine Learning Ken Clarkson Elad Hazan David Woodruff IBM Almaden Technion IBM Almaden.
Advertisements

Slow and Fast Mixing of Tempering and Swapping for the Potts Model Nayantara Bhatnagar, UC Berkeley Dana Randall, Georgia Tech.
=
On Complexity, Sampling, and -Nets and -Samples. Range Spaces A range space is a pair, where is a ground set, it’s elements called points and is a family.
An Efficient Membership-Query Algorithm for Learning DNF with Respect to the Uniform Distribution Jeffrey C. Jackson Presented By: Eitan Yaakobi Tamar.
Gibbs sampler - simple properties It’s not hard to show that this MC chain is aperiodic. Often is reversible distribution. If in addition the chain is.
Counting Algorithms for Knapsack and Related Problems 1 Raghu Meka (UT Austin, work done at MSR, SVC) Parikshit Gopalan (Microsoft Research, SVC) Adam.
1 The Monte Carlo method. 2 (0,0) (1,1) (-1,-1) (-1,1) (1,-1) 1 Z= 1 If  X 2 +Y 2  1 0 o/w (X,Y) is a point chosen uniformly at random in a 2  2 square.
1 Heat flow and a faster Algorithm to Compute the Surface Area of a Convex Body Hariharan Narayanan, University of Chicago Joint work with Mikhail Belkin,
Markov Chains 1.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Random Walks Ben Hescott CS591a1 November 18, 2002.
Entropy Rates of a Stochastic Process
BAYESIAN INFERENCE Sampling techniques
Infinite Horizon Problems
1 CE 530 Molecular Simulation Lecture 8 Markov Processes David A. Kofke Department of Chemical Engineering SUNY Buffalo
Seminar on Random Walks on Graphs 2009/2010 Ilan Ben Bassat Omri Weinstein Mixing Time – General Chains.
. PGM: Tirgul 8 Markov Chains. Stochastic Sampling  In previous class, we examined methods that use independent samples to estimate P(X = x |e ) Problem:
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
Evaluating Hypotheses
13. The Weak Law and the Strong Law of Large Numbers
Preference Analysis Joachim Giesen and Eva Schuberth May 24, 2006.
Expanders Eliyahu Kiperwasser. What is it? Expanders are graphs with no small cuts. The later gives several unique traits to such graph, such as: – High.
1 On the Computation of the Permanent Dana Moshkovitz.
Sampling and Approximate Counting for Weighted Matchings Roy Cagan.
Distributed Combinatorial Optimization
Approximating The Permanent Amit Kagan Seminar in Complexity 04/06/2001.
1 Biased card shuffling and the asymmetric exclusion process Elchanan Mossel, Microsoft Research Joint work with Itai Benjamini, Microsoft Research Noam.
Lehrstuhl für Informatik 2 Gabriella Kókai: Maschine Learning 1 Evaluating Hypotheses.
APPLICATIONS OF DIFFERENTIATION 4. In Sections 2.2 and 2.4, we investigated infinite limits and vertical asymptotes.  There, we let x approach a number.
Mixing Times of Markov Chains for Self-Organizing Lists and Biased Permutations Prateek Bhakta, Sarah Miracle, Dana Randall and Amanda Streib.
Mixing Times of Self-Organizing Lists and Biased Permutations Sarah Miracle Georgia Institute of Technology.
Monte Carlo Methods in Partial Differential Equations.
Improved results for a memory allocation problem Rob van Stee University of Karlsruhe Germany Leah Epstein University of Haifa Israel WADS 2007 WAOA 2007.
Introduction to Monte Carlo Methods D.J.C. Mackay.
Copyright © Cengage Learning. All rights reserved. CHAPTER 11 ANALYSIS OF ALGORITHM EFFICIENCY ANALYSIS OF ALGORITHM EFFICIENCY.
An Algorithmic Proof of the Lopsided Lovasz Local Lemma Nick Harvey University of British Columbia Jan Vondrak IBM Almaden TexPoint fonts used in EMF.
LIMITS AND DERIVATIVES 2. In Sections 2.2 and 2.4, we investigated infinite limits and vertical asymptotes.  There, we let x approach a number.  The.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology The Weak Law and the Strong.
Volume computation László Lovász Microsoft Research
Why is it useful to walk randomly? László Lovász Mathematical Institute Eötvös Loránd University October
An FPTAS for #Knapsack and Related Counting Problems Parikshit Gopalan Adam Klivans Raghu Meka Daniel Štefankovi č Santosh Vempala Eric Vigoda.
. CLASSES RP AND ZPP By: SARIKA PAMMI. CONTENTS:  INTRODUCTION  RP  FACTS ABOUT RP  MONTE CARLO ALGORITHM  CO-RP  ZPP  FACTS ABOUT ZPP  RELATION.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Time to Equilibrium for Finite State Markov Chain 許元春(交通大學應用數學系)
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Time Complexity of Algorithms (Asymptotic Notations)
Arrangements and Duality Motivation: Ray-Tracing Fall 2001, Lecture 9 Presented by Darius Jazayeri 10/4/01.
Seminar on random walks on graphs Lecture No. 2 Mille Gandelsman,
Diversity Loss in General Estimation of Distribution Algorithms J. L. Shapiro PPSN (Parallel Problem Solving From Nature) ’06 BISCuit 2 nd EDA Seminar.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Date: 2005/4/25 Advisor: Sy-Yen Kuo Speaker: Szu-Chi Wang.
Spatial decay of correlations and efficient methods for computing partition functions. David Gamarnik Joint work with Antar Bandyopadhyay (U of Chalmers),
Classical mathematics and new challenges László Lovász Microsoft Research One Microsoft Way, Redmond, WA Theorems and Algorithms.
Sampling algorithms and Markov chains László Lovász Microsoft Research One Microsoft Way, Redmond, WA 98052
Kevin Stevenson AST 4762/5765. What is MCMC?  Random sampling algorithm  Estimates model parameters and their uncertainty  Only samples regions of.
Theory of Computational Complexity Probability and Computing Ryosuke Sasanuma Iwama and Ito lab M1.
Introduction to Computer Simulation of Physical Systems (Lecture 10) Numerical and Monte Carlo Methods (CONTINUED) PHYS 3061.
Theory of Computational Complexity M1 Takao Inoshita Iwama & Ito Lab Graduate School of Informatics, Kyoto University.
Markov Chains Mixing Times Lecture 5
Path Coupling And Approximate Counting
Chapter 5. Optimal Matchings
Markov chain monte carlo
Enumerating Distances Using Spanners of Bounded Degree
Haim Kaplan and Uri Zwick
Hariharan Narayanan, University of Chicago Joint work with
DATABASE HISTOGRAMS E0 261 Jayant Haritsa
Presentation transcript:

The Poincaré Constant of a Random Walk in High- Dimensional Convex Bodies Ivona Bezáková Thesis Advisor: Prof. Eric Vigoda

Goal: Efficient algorithm for sampling points from high-dimensional convex bodies Approach: Random walks

Overview of the talk Motivation & History Introduction to random walks, Markov chains Definition and basic properties of ball walks Overview of the algorithm for uniform sampling Analysis of the Poincaré constant

Sampling points from convex bodies efficiently Computation of volume Sampling contingency tables [Morris] Universal portfolios [Kalai & Vempala] Convex programs [Bertsimas & Vempala] Motivation Why is computing volume difficult? Straightforward approach: - find bounding box - sample sufficiently many points from the box - compute ratio: # points in the body vs. total # points

This algorithm correctly approximates volume. Where is the catch? Ratio of the volume of the body vs. volume of the box decreases exponentially with dimension! This results in exponential running time. Note: For small dimension (2D, 3D) this algorithms works well. Goal: Find algorithm running in time polynomial in dimension.

How efficient sampling helps with volume? intersect with balls doubling in volume this defines sequence of convex bodies sample points from i- th body, compute ratio: i-th vs. (i-1)-st body return product of ratios Why is this better than the bounding box? - Volume of bodies max. doubles Result: volume of the original body vs. volume of the smallest body (a ball)

History negative results for volume computation: - cannot be approximated by deterministic polytime algorithm randomized approximation: - Dyer, Frieze, Kannan ’89 - improvements: combinations of [Applegate, Dyer, Frieze, Kannan, Lovász, Simonovits] - Kannan, Lovász, Simonovits ‘97 - #P-complete [Elekes, Bárány & Füredi, Dyer & Frieze, Khachiyan, Lawrence, Lovász & Simonovits]

Notation convex body of diameter (given by membership oracle) step-size ball unit-ball

Ball Walks Speedy Walk – next point Metropolis Walk – next point if then Problem: How to implement?

Markov Chains State space Transition distribution (likelihood of going from to ) For speedy walk we have state space = convex body if otherwise

Markov Chains 2 Stationary distribution = limiting distribution (fix-point)i.e. Mixing time For given mixing time is the expected number of steps needed to get close to the stationary distribution. Want: rapid mixing, i.e. time polynomial in and

Comparison KLS vs. this work Kannan, Lovász, Simonovits study so-called conductance for bounding mixing time. ? poly-logarithmic in We bound so-called Poincaré constant (generalization of conductance) and get mixing time cubic in spectral gap

New ideas in KLS separated analysis of speedy walk (fast mixing in principle) and Metropolis walk (efficient implementation) for volume computation: introduced isotropic position to reduce diameter of the body Why Poincaré constant? generalization might lead to better analysis through other quantities (log-Sobolev, [Frieze & Kannan, Jerrum & Son]) the same difficulty our focus: survey

Poincaré constant where Well-studied Quantities and Dirichlet form(local variance) measures decaying of variance

Well-studied Quantities 2 (Properties of Poincaré) Thm: For (lazy reversible) Markov Chain For Markov chains defined on finite state spaces the Poincaré constant equals the spectral gap. where is the distribution after steps with probability ½ stay at the same statecorresponds to symmetric chains and is the stationary distribution

Thoughts about Poincaré constant If then Thus, in this case and the chain mixes (very) rapidly. Intuitively, this corresponds to a complete graph, where we can get from any point to any other point.

Well-studied Quantities 3 Conductance equals Poincaré over indicator functions trivially Cheeger-type inequality by Jerrum and Sinclair, ‘89

Properties of Ball Walks Local conductance Ball walks: stationary distribution where reversible

From Speedy Walk to Uniform Sampling (Overview) bound Poincaré constant for speedy walk mixing time for speedy walk running time of Metropolis walk (assuming good starting distribution) obtain a good starting distribution from a sample point from the speedy distribution obtain a sample point close to the uniform distribution

From Speedy Walk to Uniform Sampling Poincaré inequality (for speedy walk): Ifthen for some dimension-independent constant Mixing time of speedy walk: For givendistribution after steps withinfrom speedy distribution (assuming reasonable starting distribution ) Thm: For (lazy reversible) Markov Chain where is the distribution after steps and is the stationary distribution

From Speedy Walk to Uniform Sampling (Overview) bound Poincaré constant for speedy walk mixing time for speedy walk running time of Metropolis walk (assuming good starting distribution) obtain a good starting distribution from a sample point from the speedy distribution obtain a sample point close to the uniform distribution

From Speedy Walk to Uniform Sampling 2 From speedy to Metropolis walk Run M. walk until speedy steps Mixing time of Metropolis walk: Ifthen we expect the total number of steps (speedy + Metropolis) (with exception )to be at most where is the average local conductance:

From Speedy Walk to Uniform Sampling (Overview) bound Poincaré constant for speedy walk mixing time for speedy walk running time of Metropolis walk (assuming good starting distribution) obtain a good starting distribution from a sample point from the speedy distribution obtain a sample point close to the uniform distribution

From Speedy Walk to Uniform Sampling 3 Obtaining a good starting distribution Let and for where Algo: Sample from according to For obtain : Run Metropolis in starting at

From Speedy Walk to Uniform Sampling 4 Good starting distribution for Metropolis walk Thm: For sufficiently small and the distribution of is within of. Expected total number of oracle calls (with exception ) is less than

From Speedy Walk to Uniform Sampling (Overview) bound Poincaré constant for speedy walk mixing time for speedy walk running time of Metropolis walk (assuming good starting distribution) obtain a good starting distribution from a sample point from the speedy distribution obtain a sample point close to the uniform distribution

From Speedy Walk to Uniform Sampling 5 From speedy distribution to the uniform distribution Algo: Shrink Sample from until Return Thm: If and sufficiently small then the distribution of is away from the uniform distribution. Expected number of samples needed.

From Speedy Walk to Uniform Sampling (Overview) bound Poincaré constant for speedy walk mixing time for speedy walk running time of Metropolis walk (assuming good starting distribution) obtain a good starting distribution from a sample point from the speedy distribution obtain a sample point close to the uniform distribution

Proof of the Poincaré Inequality Restricted variance, Dirichlet form, expected value Poincaré inequality (for speedy walk): Ifthen for any function for some dimension-independent constant where

Idea of the proof: For a sufficiently small set such that does not vary much within Assuming Poincaré does not hold, we find a set contradicting the above Find needle-like s.t. wlog and Chop to obtain desired set

Needle-like Body Eliminate dimensions one by one (inductively) assume has fat dimensions and while projection of onto two fat dim. there exists a point s.t. any line through cuts into appx. half take hyperplane s.t. at least one of these must be true or

Shrinking Last Dimension Goal: find s.t. last dim. of is and where is a constant (dependent on ) How? Chop into Ideally ? But

Assumption Idea: relate towhere We get where and Next goal: bound

Chopping of What do we need? does not vary much within, i.e. for any let width of is at most We will show that this chopping allows us to bound appropriately

Properties of local conductance is concave over From Brunn-Minkowski Thm: is Lipschitz over : For any Implications for the The width of increases, then it is (full width) and then it decreases For sufficiently small the width of any is at least

From Dinghas’ Thm: For the middle section is convex. Now we can split into several sums and estimate them separately Thus where is the number of slabs in the middle section What to do outside the middle section?

In the left section, the increase exponentially This allows us to bound We obtain similar bounds for other parts of the sum, putting them together we get We wanted Thus we proved the Poincaré inequality.

THANK YOU