Volume distortion for subsets of R n James R. Lee Institute for Advanced Study & University of Washington Symposium on Computational Geometry, 2006; Sedona,

Slides:



Advertisements
Similar presentations
K The Common Core State Standards in Mathematics © Copyright 2011 Institute for Mathematics and Education Welcome to a clickable.
Advertisements

Vertex sparsifiers: New results from old techniques (and some open questions) Robert Krauthgamer (Weizmann Institute) Joint work with Matthias Englert,
Distance Preserving Embeddings of Low-Dimensional Manifolds Nakul Verma UC San Diego.
TexPoint fonts used in EMF.
Comments We consider in this topic a large class of related problems that deal with proximity of points in the plane. We will: 1.Define some proximity.
Voronoi Diagrams in n· 2 O(√lglg n ) Time Timothy M. ChanMihai Pătraşcu STOC’07.
Trees and Markov convexity James R. Lee Institute for Advanced Study [ with Assaf Naor and Yuval Peres ] RdRd x y.
Metric Embeddings with Relaxed Guarantees Hubert Chan Joint work with Kedar Dhamdhere, Anupam Gupta, Jon Kleinberg, Aleksandrs Slivkins.
Metric embeddings, graph expansion, and high-dimensional convex geometry James R. Lee Institute for Advanced Study.
Geometric embeddings and graph expansion James R. Lee Institute for Advanced Study (Princeton) University of Washington (Seattle)
Interchanging distance and capacity in probabilistic mappings Uriel Feige Weizmann Institute.
A Nonlinear Approach to Dimension Reduction Lee-Ad Gottlieb Weizmann Institute of Science Joint work with Robert Krauthgamer TexPoint fonts used in EMF.
Distance Scales, Embeddings, and Metrics of Negative Type By James R. Lee Presented by Andy Drucker Mar. 8, 2007 CSE 254: Metric Embeddings.
Navigating Nets: Simple algorithms for proximity search Robert Krauthgamer (IBM Almaden) Joint work with James R. Lee (UC Berkeley)
With thanks to Zhijun Wu An introduction to the algorithmic problems of Distance Geometry.
“Random Projections on Smooth Manifolds” -A short summary
On Approximating the Average Distance Between Points Kfir Barhum, Oded Goldreich and Adi Shraibman Weizmann Institute of Science.
Proximity algorithms for nearly-doubling spaces Lee-Ad Gottlieb Robert Krauthgamer Weizmann Institute TexPoint fonts used in EMF. Read the TexPoint manual.
1 Optimization problems such as MAXSAT, MIN NODE COVER, MAX INDEPENDENT SET, MAX CLIQUE, MIN SET COVER, TSP, KNAPSACK, BINPACKING do not have a polynomial.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Spring, 2003 Review Lecture Tuesday, 5/6/03.
Network Design Adam Meyerson Carnegie-Mellon University.
Probabilistic Methods in Coding Theory: Asymmetric Covering Codes Joshua N. Cooper UCSD Dept. of Mathematics Robert B. Ellis Texas A&M Dept. of Mathematics.
An Approximation Algorithm for Requirement cut on graphs Viswanath Nagarajan Joint work with R. Ravi.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2002 Review Lecture Tuesday, 12/10/02.
Lower Bounds on the Distortion of Embedding Finite Metric Spaces in Graphs Y. Rabinovich R. Raz DCG 19 (1998) Iris Reinbacher COMP 670P
Efficient Nearest-Neighbor Search in Large Sets of Protein Conformations Fabian Schwarzer Itay Lotan.
Dimensionality Reduction
Greedy Routing with Bounded Stretch Roland Flury, Roger Wattenhofer (ETH Zurich), Sriram Pemmaraju (Iowa University) Published at IEEE Infocom 2009 Introduction.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
Dimensionality Reduction. Multimedia DBs Many multimedia applications require efficient indexing in high-dimensions (time-series, images and videos, etc)
Embedding and Sketching Alexandr Andoni (MSR). Definition by example  Problem: Compute the diameter of a set S, of size n, living in d-dimensional ℓ.
Random Projections of Signal Manifolds Michael Wakin and Richard Baraniuk Random Projections for Manifold Learning Chinmay Hegde, Michael Wakin and Richard.
cover times, blanket times, and majorizing measures Jian Ding U. C. Berkeley James R. Lee University of Washington Yuval Peres Microsoft Research TexPoint.
Distance scales, embeddings, and efficient relaxations of the cut cone James R. Lee University of California, Berkeley.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Footer Here1 Feature Selection Copyright, 1996 © Dale Carnegie & Associates, Inc. David Mount For CMSC 828K: Algorithms and Data Structures for Information.
Algorithms on negatively curved spaces James R. Lee University of Washington Robert Krauthgamer IBM Research (Almaden) TexPoint fonts used in EMF. Read.
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Geometric Problems in High Dimensions: Sketching Piotr Indyk.
Advanced Algorithm Design and Analysis (Lecture 13) SW5 fall 2004 Simonas Šaltenis E1-215b
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
13 th Nov Geometry of Graphs and It’s Applications Suijt P Gujar. Topics in Approximation Algorithms Instructor : T Kavitha.
Cover times, blanket times, and the GFF Jian Ding Berkeley-Stanford-Chicago James R. Lee University of Washington Yuval Peres Microsoft Research.
BCS547 Neural Decoding.
Geometric Problems in High Dimensions: Sketching Piotr Indyk.
Embeddings, flow, and cuts: an introduction University of Washington James R. Lee.
Doubling Dimension: a short survey Anupam Gupta Carnegie Mellon University Barriers in Computational Complexity II, CCI, Princeton.
Lower Bounds for Embedding Edit Distance into Normed Spaces A. Andoni, M. Deza, A. Gupta, P. Indyk, S. Raskhodnikova.
Sampling Theory and Some Important Sampling Distributions.
What is a metric embedding?Embedding ultrametrics into R d An embedding of an input metric space into a host metric space is a mapping that sends each.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2001 Review Lecture Tuesday, 12/11/01.
Multi-way spectral partitioning and higher-order Cheeger inequalities University of Washington James R. Lee Stanford University Luca Trevisan Shayan Oveis.
Modeling K The Common Core State Standards in Mathematics Geometry Measurement and Data The Number System Number and Operations.
1 Approximation Algorithms for Low- Distortion Embeddings into Low- Dimensional Spaces Badoiu et al. (SODA 2005) Presented by: Ethan Phelps-Goodman Atri.
Coarse Differentiation and Planar Multiflows
Dimension reduction for finite trees in L1
Optimization problems such as
Generalized Sparsest Cut and Embeddings of Negative-Type Metrics
Haim Kaplan and Uri Zwick
Ultra-low-dimensional embeddings of doubling metrics
Sublinear Algorithmic Tools 3
Greedy Routing with Bounded Stretch
Lecture 16: Earth-Mover Distance
Randomized Algorithms CS648
Near-Optimal (Euclidean) Metric Compression
Light Spanners for Snowflake Metrics
Metric Methods and Approximation Algorithms
Embedding Metrics into Geometric Spaces
Lecture 15: Least Square Regression Metric Embeddings
Presentation transcript:

volume distortion for subsets of R n James R. Lee Institute for Advanced Study & University of Washington Symposium on Computational Geometry, 2006; Sedona, AZ

graph bandwidth Given a graph G=(V,E), we seek a permutation  : V ! {1,2,...,n}. The bandwidth of  is The bandwidth of G is bw(G) = min  bw(  ). 11 j=uj=u nn k=vk=v Goal: Efficiently compute an ordering  for which bw(  ) ¼ bw(G).

embeddings & projections 1 ) Embed G into a Euclidean space R k (preserving distances) 2 ) Project onto a random line and take the induced linear ordering.

embeddings & projections Analysis: Count the number of points that fall into an interval, and use this to upper bound the stretch. Problem: Only have control on the expectation, but there could be costly correlations...

embeddings & projections Analysis: Count the number of points that fall into an interval, and use this to upper bound the stretch. Problem: Only have control on the expectation, but there could be costly correlations... Feige gave an example where this approach will take a graph of bandwidth 2, and with high probability yield a solution of bandwidth

embeddings & projections Analysis: Count the number of points that fall into an interval, and use this to upper bound the stretch. Feige gave an example where this approach will take a graph of bandwidth 2, and with high probability yield a solution of bandwidth

embeddings & projections Volume of a set of points controls the probability they project close together. Conditioned on the projection of the three red points, the projection of the blue point still varies proportional to the distance to the affine hull of the red points (false, but essentially true in high dimensions)

volume distortion This leads to a new notion of embedding [Feige 00], one which tries to maximize the volume of e.g. all triangles in the image. non-expansive We would like to get as close as possible to the best possible volume for all the triples in our metric space (e.g. shortest path on our graph)

higher-dimensional distortion Given a metric space (X,d), a number k, and a non-expansive mapping f : X ! R m, we define the k-dimensional distortion of f as the smallest number D such that... (note: 1 -dimensional distortion recovers the “standard” notion)

previous results General n-point metric spaces [Feige 97] [Rao 99] [Krauthgamer-L-Mendel-Naor 04] Subsets of Euclidean spaces (important to analyze bandwidth SDP) [Rao 99] [Krauthgamer-Linial-Magen 03 Dunagan-Vempala 01] New results: (k = 3)

main result Given an n-point subset X µ R n, there exists a non-expansive mapping f : X ! R n such that... This embedding maximizes the volume of every k-point subset within factor ¼ (log n) k/2.

construction of the embedding Three phases 1)Randomized reduction to a collection of polylog(n) easier problems Random partitions, random sampling, gluing via smooth bump functions, measured descent,... [KLMN, L, ALN]

construction of the embedding 2) Reduction to a continuous problem in the “right” dimension Dimension reduction (Johnson-Lindenstrauss) and Kirszbraun’s extension theorem RkRk Lipschitz extension problem: Given S µ X and a non-expansive map f : S ! R k, does there exist an non-expansive extension f : X ! R k ? Answer: Yes if X is a subset of Euclidean space

kirszbraun’s theorem

construction of the embedding 3) Solution of the continuous problem If you think of F d (x) as being a real-valued random variable for every x, then we are saying that (standard deviation)

construction of the embedding If you think of F d (x) as being a real-valued random variable for every x, then we are saying that x Q y

open problems – remove the O(log log n) terms (here and SparsestCut ) / simplify analysis – improve the approximation ratio for bandwidth best known is ¼ O(log n) 3 [Feige, Dunagan-Vempala] best known for trees is O(log n) 2.5 [Gupta] conjectured optimal bound: O(log n)