Proximity algorithms for nearly-doubling spaces Lee-Ad Gottlieb Robert Krauthgamer Weizmann Institute TexPoint fonts used in EMF. Read the TexPoint manual.

Slides:



Advertisements
Similar presentations
Efficient classification for metric data Lee-Ad GottliebWeizmann Institute Aryeh KontorovichBen Gurion U. Robert KrauthgamerWeizmann Institute TexPoint.
Advertisements

The Primal-Dual Method: Steiner Forest TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A AA A A A AA A A.
Weighted Matching-Algorithms, Hamiltonian Cycles and TSP
Min-Max Relations, Hall’s Theorem, and Matching-Algorithms Graphs & Algorithms Lecture 5 TexPoint fonts used in EMF. Read the TexPoint manual before you.
WSPD Applications.
C&O 355 Lecture 23 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A A A A A.
Approximation Algorithms
Generalization and Specialization of Kernelization Daniel Lokshtanov.
Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
Approximation, Chance and Networks Lecture Notes BISS 2005, Bertinoro March Alessandro Panconesi University La Sapienza of Rome.
Doubling dimension and the traveling salesman problem Yair BartalHebrew University Lee-Ad GottliebHebrew University Robert KrauthgamerWeizmann Institute.
Approximation Algorithms for TSP
Approximating the Domatic Number Feige, Halldorsson, Kortsarz, Srinivasan ACM Symp. on Theory of Computing, pages , 2000.
Approximation Algorithms Chapter 5: k-center. Overview n Main issue: Parametric pruning –Technique for approximation algorithms n 2-approx. algorithm.
A polylogarithmic approximation of the minimum bisection Robert Krauthgamer The Hebrew University Joint work with Uri Feige.
A Nonlinear Approach to Dimension Reduction Lee-Ad Gottlieb Weizmann Institute of Science Joint work with Robert Krauthgamer TexPoint fonts used in EMF.
Optimization of Pearl’s Method of Conditioning and Greedy-Like Approximation Algorithm for the Vertex Feedback Set Problem Authors: Ann Becker and Dan.
Navigating Nets: Simple algorithms for proximity search Robert Krauthgamer (IBM Almaden) Joint work with James R. Lee (UC Berkeley)
Approximating Maximum Edge Coloring in Multigraphs
Complexity 15-1 Complexity Andrei Bulatov Hierarchy Theorem.
Routing, Anycast, and Multicast for Mesh and Sensor Networks Roland Flury Roger Wattenhofer RAM Distributed Computing Group.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey University of Waterloo Department of Combinatorics and Optimization Joint.
Efficient classification for metric data Lee-Ad GottliebHebrew U. Aryeh KontorovichBen Gurion U. Robert KrauthgamerWeizmann Institute TexPoint fonts used.
Approximation Algorithms: Combinatorial Approaches Lecture 13: March 2.
CPSC 689: Discrete Algorithms for Mobile and Wireless Systems Spring 2009 Prof. Jennifer Welch.
A general approximation technique for constrained forest problems Michael X. Goemans & David P. Williamson Presented by: Yonatan Elhanani & Yuval Cohen.
1 CSE 417: Algorithms and Computational Complexity Winter 2001 Lecture 23 Instructor: Paul Beame.
Computational Complexity, Physical Mapping III + Perl CIS 667 March 4, 2004.
Matrix sparsification and the sparse null space problem Lee-Ad GottliebWeizmann Institute Tyler NeylonBynomial Inc. TexPoint fonts used in EMF. Read the.
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
Steiner trees Algorithms and Networks. Steiner Trees2 Today Steiner trees: what and why? NP-completeness Approximation algorithms Preprocessing.
Introduction Outline The Problem Domain Network Design Spanning Trees Steiner Trees Triangulation Technique Spanners Spanners Application Simple Greedy.
Finding a maximum independent set in a sparse random graph Uriel Feige and Eran Ofek.
1 Refined Search Tree Technique for Dominating Set on Planar Graphs Jochen Alber, Hongbing Fan, Michael R. Fellows, Henning Fernau, Rolf Niedermeier, Fran.
Outline Introduction The hardness result The approximation algorithm.
Efficient Regression in Metric Spaces via Approximate Lipschitz Extension Lee-Ad GottliebAriel University Aryeh KontorovichBen-Gurion University Robert.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Simple and Improved Parameterized Algorithms for Multiterminal Cuts Mingyu Xiao The Chinese University of Hong Kong Hong Kong SAR, CHINA CSR 2008 Presentation,
The Best Algorithms are Randomized Algorithms N. Harvey C&O Dept TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A AAAA.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
A Clustering Algorithm based on Graph Connectivity Balakrishna Thiagarajan Computer Science and Engineering State University of New York at Buffalo.
1 Sublinear Algorithms Lecture 1 Sofya Raskhodnikova Penn State University TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Fast, precise and dynamic distance queries Yair BartalHebrew U. Lee-Ad GottliebWeizmann → Hebrew U. Liam RodittyBar Ilan Tsvi KopelowitzBar Ilan → Weizmann.
Discrete Structures Lecture 12: Trees Ji Yanyan United International College Thanks to Professor Michael Hvidsten.
Computational Geometry Piyush Kumar (Lecture 10: Point Location) Welcome to CIS5930.
An optimal dynamic spanner for points residing in doubling metric spaces Lee-Ad Gottlieb NYU Weizmann Liam Roditty Weizmann.
Geometric Problems in High Dimensions: Sketching Piotr Indyk.
Presenter : Kuang-Jui Hsu Date : 2011/3/24(Thur.).
Approximation Algorithms for TSP Tsvi Kopelowitz 1.
A light metric spanner Lee-Ad Gottlieb. Graph spanners A spanner for graph G is a subgraph H ◦ H contains vertices, subset of edges of G Some qualities.
The full Steiner tree problem Theoretical Computer Science 306 (2003) C. L. Lu, C. Y. Tang, R. C. T. Lee Reporter: Cheng-Chung Li 2004/06/28.
1 Approximation algorithms Algorithms and Networks 2015/2016 Hans L. Bodlaender Johan M. M. van Rooij TexPoint fonts used in EMF. Read the TexPoint manual.
Approximating Buy-at-Bulk and Shallow-Light k-Steiner Trees Mohammad T. Hajiaghayi (CMU) Guy Kortsarz (Rutgers) Mohammad R. Salavatipour (U. Alberta) Presented.
On the Impossibility of Dimension Reduction for Doubling Subsets of L p Yair Bartal Lee-Ad Gottlieb Ofer Neiman.
Final Review Chris and Virginia. Overview One big multi-part question. (Likely to be on data structures) Many small questions. (Similar to those in midterm.
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. Fast.
1 Approximation Algorithms for Low- Distortion Embeddings into Low- Dimensional Spaces Badoiu et al. (SODA 2005) Presented by: Ethan Phelps-Goodman Atri.
Introduction Wireless Ad-Hoc Network  Set of transceivers communicating by radio.
Nearly optimal classification for semimetrics
8.3.2 Constant Distance Approximations
Approximation Algorithms for TSP
Enumerating Distances Using Spanners of Bounded Degree
Randomized Algorithms CS648
Yair Bartal Lee-Ad Gottlieb Hebrew U. Ariel University
Coverage Approximation Algorithms
On the effect of randomness on planted 3-coloring models
Introduction Wireless Ad-Hoc Network
The Full Steiner tree problem Part Two
Clustering.
Presentation transcript:

Proximity algorithms for nearly-doubling spaces Lee-Ad Gottlieb Robert Krauthgamer Weizmann Institute TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAA

Proximity algorithms for nearly-doubling spaces 2 Proximity problems In arbitrary metric space, some proximity problems are hard  For example, the nearest neighbor search problem requires Θ(n) time The doubling dimension parameterizes the “bad” case… q ~1

Proximity algorithms for nearly-doubling spaces 3 Doubling Dimension Definition: Ball B(x,r) = all points within distance r from x. The doubling constant (of a metric M) is the minimum value ¸  such that every ball can be covered by ¸ balls of half the radius  First used by [Ass-83], algorithmically by [Cla-97].  The doubling dimension is dim(M)=log ¸ (M) [GKL-03]  A metric is doubling if its doubling dimension is constant Packing property of doubling spaces  A set with diameter D and min. inter-point distance a, contains at most (D/a) O(log ¸ ) points Here ≤7.

Proximity algorithms for nearly-doubling spaces 4 Applications In the past few years, many algorithmic tasks have been analyzed via the doubling dimension  For example, approximate nearest neighbor search can be executed in time ¸ O(1) log n Some other algorithms analyzed via the doubling dimension  Nearest neighbor search [KL-04, BKL-06, CG-06]  Clustering [Tal-04, ABS-08, FM-10]  Spanner construction [GGN-06, CG-06, DPP-06, GR-08]  Routing [KSW-04, Sil-05, AGGM-06, KRXY-07, KRX-08]  Travelling Salesperson [Tal-04]  Machine learning [BLL-09, GKK-10] Message: This is an active line of research…

Proximity algorithms for nearly-doubling spaces 5 Problem Most algorithms developed for doubling spaces are not robust  Algorithmic guarantees don’t hold for nearly-doubling spaces  If a small fraction of the working set possesses high doubling dimension, algorithmic performance degrades. This problem motivates the following key task  Given an n-point set S and target dimension d*  Remove from S the fewest number of points so that the remaining set has doubling dimension at most d*

Proximity algorithms for nearly-doubling spaces 6 Two paradigms How can removing a few “bad” points help? Two models: 1. Ignore the bad points  Outlier detection. [GHPT-05] cluster based on similarity, seek a large subset with low intrinsic dimension.  Algorithms with slack. Throw bad points into the slack [KRXY-07] gave a routing algorithm with guarantees for most of the input points. [FM-10] gave a kinetic clustering algorithm for most of the input points. [GKK-10] gave a machine learning algorithm – small subset doesn’t interfere with learning

Proximity algorithms for nearly-doubling spaces 7 Two paradigms How can removing a few “bad” points help? Two models: 2. Tailor a different algorithm for the bad points  Example: Spanner construction. A spanner is an edge subset of the full graph Good points: Low doubling dimensionsparse spanner with nice properties (low stretch and degree) Bad points: Take the full graph If the number of bad points is O(n.5 ), we have a spanner with O(n) edges

Proximity algorithms for nearly-doubling spaces 8 Results Recall our key problem  Given an n-point set S and target dimension d*  Remove from S the fewest number of points so that the remaining set has doubling dimension at most d* This problem is NP-hard  Even determining the doubling dimension of a point set exactly is NP- hard! Proof on the next slide But the doubling dimension can be approximated within a constant factor… Our contribution: bicriteria approximation algorithm  In time 2 O(d*) n 3, we remove a number of points arbitrarily close to optimal, while achieving doubling dimension 4d* + O(1)  We can also achieve near-linear runtime, at the cost of slightly higher dimension

Proximity algorithms for nearly-doubling spaces 9 Warm up Lemma: It is NP-hard to determine the doubling dimension of a set S  Reduction: from vertex cover with bounded degree Δ = n ½. the size of any vertex cover is at least n ½.  Construction: A set S of n points corresponding to the vertex set V. Let d(u,v) = ½ if the cor. vertices are connected by an edge Let d(u,v) = 1if the cor. vertices aren’t connected  Analysis: Any subset of S found in a ball of radius ½ has at most n ½ points - degree of original graph S is a ball of radius 1. The minimum covering of all of S with balls of radius ½ is equal to the minimum vertex cover of V. Note: reduction preserves hardness of approximation Corollary: It is NP-hard to determine if removing k points from S can leave a set with doubling dimension d*.  So our problem is hard as well. ½ ½ 1

Proximity algorithms for nearly-doubling spaces 10 Bicriteria algorithm Recall that he doubling constant (of a metric M) is  the minimum value ¸  such that every r-radius ball can be covered by ¸ balls of half the radius Define the related notion of density constant as  the minimum value  >0 such that every r-radius ball contains at most  points at mutual interpoint distance r/2  Nice property: The density constant can only decrease under the removal of points, unlike the doubling constant. We can show that  √  (S) ≤ ¸ (S) ≤  (S)  it’s NP-hard to compute the density constant (ratio-preserving reduction from independent set) =2,  =3

Proximity algorithms for nearly-doubling spaces 11 Bicriteria algorithm We will give a bicriteria algorithm for the density constant. Problem statement:  Given an n-point set S and target density constant  *  Remove from S the fewest number of points so that the remaining set has density constant at most  * A bicriteria algorithm for the density constant is itself a bicriteria algorithm for the doubling constant  within a quadratic factor

Proximity algorithms for nearly-doubling spaces 12 Witness set Given a set S, a subset S’ is a witness set for the density constant if  All points are at interpoint distance at least r/2  Note that S’ is a concise proof that the density constant of S is at least |S’| Theorem: Fix a value  ’<  (S). A witness set of S of size at least √  ‘ can be found in time 2 O(  *) n 3  Proof outline:  For each point p and radius r define the r-ball of p.  Greedily cover all points in the r-ball with disjoint balls of radius r/2.  Then cover all points in each r/2 ball with disjoint balls of radius r/4.  Since there exists in S a witness set of size  (S), there exists a p and r so that either there are √  (S) r/2 balls, and these form a witness set, or one r/2 ball covers √  (S) r/4 balls, and these form a witness set.

Proximity algorithms for nearly-doubling spaces 13 Bicriteria algorithm Recall our problem  Given an n-point set S and target density constant  *  Remove from S the fewest number of points so that the remaining set has density constant at most  * Our bricriteria solution:  Let k be the true answer (the minimum number of points that must be removed).  We remove k c/(c-1) points and the remaining set has density constant c 2  * 2

Proximity algorithms for nearly-doubling spaces 14 Bicriteria algorithm Algorithm  Run the subroutine to identify a witness set of size at least c  *  Remove it  Repeat Analysis  The density constant of the resulting set is not greater than c 2  * 2 since we terminated without finding a witness set of size at least c  *  Every time a witness set of size w>c  * is removed by our algorithm, the optimal algorithm must remove at least w-  * points or else the true solution would have density constant greater than  *  It follows that are algorithm removes k w/(w-  *) < kc/(c-1) points

Proximity algorithms for nearly-doubling spaces 15 Conclusion We conclude that there exists a bicriteria algorithm for the density constant  We remove k c/(c-1) points and the remaining set has density constant c 2  * 2 It follows that there exists a bricriteria algorithm for the doubling constant  We remove k c/(c-1) points and the remaining set has doubling constant c 4 ¸ * 4