Matroid Bases and Matrix Concentration

Slides:



Advertisements
Similar presentations
The Primal-Dual Method: Steiner Forest TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A AA A A A AA A A.
Advertisements

1 LP, extended maxflow, TRW OR: How to understand Vladimirs most recent work Ramin Zabih Cornell University.
C&O 355 Lecture 15 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A A A A A A A.
Weighted Matching-Algorithms, Hamiltonian Cycles and TSP
Vertex sparsifiers: New results from old techniques (and some open questions) Robert Krauthgamer (Weizmann Institute) Joint work with Matthias Englert,
Great Theoretical Ideas in Computer Science
Submodular Set Function Maximization via the Multilinear Relaxation & Dependent Rounding Chandra Chekuri Univ. of Illinois, Urbana-Champaign.
Min-Max Relations, Hall’s Theorem, and Matching-Algorithms Graphs & Algorithms Lecture 5 TexPoint fonts used in EMF. Read the TexPoint manual before you.
The Solution of the Kadison-Singer Problem Adam Marcus (Crisply, Yale) Daniel Spielman (Yale) Nikhil Srivastava (MSR/Berkeley) IAS, Nov 5, 2014.
C&O 355 Mathematical Programming Fall 2010 Lecture 6 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A.
Sparse Approximations
A. S. Morse Yale University University of Minnesota June 4, 2014 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A.
C&O 355 Lecture 23 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A A A A A.
Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
C&O 355 Mathematical Programming Fall 2010 Lecture 22 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
C&O 355 Mathematical Programming Fall 2010 Lecture 21 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
Matrix Concentration Nick Harvey University of British Columbia TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Dependent Randomized Rounding in Matroid Polytopes (& Related Results) Chandra Chekuri Jan VondrakRico Zenklusen Univ. of Illinois IBM ResearchMIT.
Approximation Algorithms for TSP Mohit Singh McGill University (in lieu of Amin Saberi)
Analysis of Network Diffusion and Distributed Network Algorithms Rajmohan Rajaraman Northeastern University, Boston May 2012 Chennai Network Optimization.
A Randomized Polynomial-Time Simplex Algorithm for Linear Programming Daniel A. Spielman, Yale Joint work with Jonathan Kelner, M.I.T.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo Department of Combinatorics and Optimization Joint work with Isaac.
Graph Sparsifiers: A Survey Nick Harvey Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman,
Graph Sparsifiers: A Survey Nick Harvey UBC Based on work by: Batson, Benczur, de Carli Silva, Fung, Hariharan, Harvey, Karger, Panigrahi, Sato, Spielman,
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey University of Waterloo Department of Combinatorics and Optimization Joint.
Graph Sparsifiers by Edge-Connectivity and Random Spanning Trees Nick Harvey U. Waterloo C&O Joint work with Isaac Fung TexPoint fonts used in EMF. Read.
Max-Min Fair Allocation of Indivisible Goods Amin Saberi Stanford University Joint work with Arash Asadpour TexPoint fonts used in EMF. Read the TexPoint.
Approximation Algorithms
1 Optimization problems such as MAXSAT, MIN NODE COVER, MAX INDEPENDENT SET, MAX CLIQUE, MIN SET COVER, TSP, KNAPSACK, BINPACKING do not have a polynomial.
The Theory of NP-Completeness
Message Passing Algorithms for Optimization
Is the following graph Hamiltonian- connected from vertex v? a). Yes b). No c). I have absolutely no idea v.
cover times, blanket times, and majorizing measures Jian Ding U. C. Berkeley James R. Lee University of Washington Yuval Peres Microsoft Research TexPoint.
C&O 355 Lecture 2 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A.
Approximating the MST Weight in Sublinear Time Bernard Chazelle (Princeton) Ronitt Rubinfeld (NEC) Luca Trevisan (U.C. Berkeley)
Theory of Computing Lecture 10 MAS 714 Hartmut Klauck.
Graph Sparsifiers Nick Harvey University of British Columbia Based on joint work with Isaac Fung, and independent work of Ramesh Hariharan & Debmalya Panigrahi.
C&O 355 Mathematical Programming Fall 2010 Lecture 19 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
Approximating Minimum Bounded Degree Spanning Tree (MBDST) Mohit Singh and Lap Chi Lau “Approximating Minimum Bounded DegreeApproximating Minimum Bounded.
An Algorithmic Proof of the Lopsided Lovasz Local Lemma Nick Harvey University of British Columbia Jan Vondrak IBM Almaden TexPoint fonts used in EMF.
Ran El-Yaniv and Dmitry Pechyony Technion – Israel Institute of Technology, Haifa, Israel Transductive Rademacher Complexity and its Applications.
Graph Sparsifiers Nick Harvey Joint work with Isaac Fung TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
An algorithmic proof of the Lovasz Local Lemma via resampling oracles Jan Vondrak IBM Almaden TexPoint fonts used in EMF. Read the TexPoint manual before.
C&O 355 Mathematical Programming Fall 2010 Lecture 18 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
Spectrally Thin Trees Nick Harvey University of British Columbia Joint work with Neil Olver (MIT  Vrije Universiteit) TexPoint fonts used in EMF. Read.
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
C&O 355 Mathematical Programming Fall 2010 Lecture 16 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A.
CPSC 536N Sparse Approximations Winter 2013 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAA.
C&O 355 Lecture 24 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A A A A A A A.
Effective-Resistance-Reducing Flows, Spectrally Thin Trees, and ATSP Nima Anari UC Berkeley Shayan Oveis Gharan Univ of Washington.
An algorithmic proof of the Lovasz Local Lemma via resampling oracles Jan Vondrak IBM Almaden TexPoint fonts used in EMF. Read the TexPoint manual before.
Generating Random Spanning Trees via Fast Matrix Multiplication Keyulu Xu University of British Columbia Joint work with Nick Harvey TexPoint fonts used.
An algorithmic proof of the Lovasz Local Lemma via resampling oracles Jan Vondrak IBM Almaden TexPoint fonts used in EMF. Read the TexPoint manual before.
Laplacian Matrices of Graphs: Algorithms and Applications ICML, June 21, 2016 Daniel A. Spielman.
Theory of Computational Complexity Probability and Computing Chapter Hikaru Inada Iwama and Ito lab M1.
Lap Chi Lau we will only use slides 4 to 19
Optimization problems such as
Topics in Algorithms Lap Chi Lau.
Approximating the MST Weight in Sublinear Time
Joint work with Frans Schalekamp and Anke van Zuylen
Density Independent Algorithms for Sparsifying
Structural Properties of Low Threshold Rank Graphs
A Constant-Factor Approximation Algorithm for the Asymmetric Traveling Salesman Problem Ola Svensson, Jakub Tarnawski and László A. Végh.
CIS 700: “algorithms for Big Data”
Matrix Martingales in Randomized Numerical Linear Algebra
Sublinear Algorihms for Big Data
Richard Anderson Lecture 28 Coping with NP-Completeness
CSCI B609: “Foundations of Data Science”
András Sebő and Anke van Zuylen
On Solving Linear Systems in Sublinear Time
Presentation transcript:

Matroid Bases and Matrix Concentration Nick Harvey University of British Columbia Joint work with Neil Olver (Vrije Universiteit) TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAA

Scalar concentration inequalities Theorem: [Chernoff / Hoeffding Bound] Let Y1,…,Ym be independent, non-negative scalar random variables. Let Y=i Yi and ¹=E [ Y ]. Suppose Yi · 1 a.s. Then

Scalar concentration inequalities Theorem: [Panconesi-Srinivasan ‘92, Dubhashi-Ranjan ‘96, etc.] Let Y1,…,Ym be negatively dependent, non-negative scalar rvs. Let Y=i Yi and ¹=E [ Y ]. Suppose Yi · 1 a.s. Then Negative cylinder dependence: Yi 2 {0,1}, Stronger notions: negative association, determinantal distributions, strongly Rayleigh measures, etc.

Matrix concentration inequalities Theorem: [Tropp ‘12, etc.] Let Y1,…,Ym be independent, PSD matrices of size nxn. Let Y=i Yi and M=E [ Y ]. Suppose ¹¢Yi ¹ M a.s. Then

Extensions of Chernoff Bounds Independent Negative Dependent Scalars Chernoff-Hoeffding Panconesi-Srinivasan, etc. Matrices Tropp, etc. ? This talk: a special case of the missing common generalization, where the negatively dependent distribution is a certain random walk in a matroid base polytope.

Knowing that e2T decreases probability that f2T Negative Dependence Arises in many natural scenarios. Random spanning trees: Let Ye indicate if edge e is in tree. Knowing that e2T decreases probability that f2T e f

Negative Dependence Arises in many natural scenarios. Random spanning trees: Let Ye indicate if edge e is in tree. Balls and bins: Let Yi be number of balls in bin i. Sampling without replacement, random permutations, random cluster models, etc.

Thin trees S S Cut ±(S) = { edge st : s2S, tS } A spanning tree T is ®-thin if |±T(S)| · ®¢|±G(S)| 8S Global connectivity: K = min {|±G(S)| : ;(S(V } Conjecture [Goddyn ’80s]: Every n-vertex graph has an ®-thin tree with ®=O(1/K). Would have deep consequences in graph theory. Let me begin the talk by motivating what sparsification is and why it’s important. Sparsification is a concept that’s familiar from our daily lives: you want to take some object that’s dense, and replace by an object that’s sparser, but almost as good as the original object. Here’s an example I learned about when I bought my first house a few years ago: floor joists used to be made of solid wood, but now they’re “engineered”, often with a truss structure, often with other materials. Another familiar example is image compression: we express the image in a particular basis (perhaps a wavelet basis), then keep the most important components and throw away the less important ones.

Thin trees S S Cut ±(S) = { edge st : s2S, tS } A spanning tree T is ®-thin if |±T(S)| · ®¢|±G(S)| 8S Global connectivity: K = min {|±G(S)| : ;(S(V } Theorem [Asadpour et al ‘10]: Every n-vertex graph has an ®-thin spanning tree with ®= . Uses negative dependence and Chernoff bounds. Let me begin the talk by motivating what sparsification is and why it’s important. Sparsification is a concept that’s familiar from our daily lives: you want to take some object that’s dense, and replace by an object that’s sparser, but almost as good as the original object. Here’s an example I learned about when I bought my first house a few years ago: floor joists used to be made of solid wood, but now they’re “engineered”, often with a truss structure, often with other materials. Another familiar example is image compression: we express the image in a particular basis (perhaps a wavelet basis), then keep the most important components and throw away the less important ones.

Asymmetric Traveling Salesman Problem [Julia Robinson, 1949] Let D=(V,E,w) be a weighted, directed graph. Goal: Find a tour sequence v1,v2,…,vk=v1 of vertices that visits every vertex in V at least once, has vivi+12E for every i, and minimizes total weight §1·i·k w(vivi+1).

Asymmetric Traveling Salesman Problem [Julia Robinson, 1949] Let D=(V,E,w) be a weighted, directed graph. Goal: Find a tour sequence v1,v2,…,vk=v1 of vertices that visits every vertex in V at least once, has vivi+12E for every i, and minimizes total weight §1·i·k w(vivi+1). Reduction [Oveis Gharan, Saberi ‘11]: If you can efficiently find an ®/K-thin spanning tree in any n-vertex graph, then you can find a tour whose weight is within O(®) of optimal.

Graph Laplacians Laplacian of edge bc Lbc = a c d b a b c d a 1 -1 b c 1 -1 b Let me begin the talk by motivating what sparsification is and why it’s important. Sparsification is a concept that’s familiar from our daily lives: you want to take some object that’s dense, and replace by an object that’s sparser, but almost as good as the original object. Here’s an example I learned about when I bought my first house a few years ago: floor joists used to be made of solid wood, but now they’re “engineered”, often with a truss structure, often with other materials. Another familiar example is image compression: we express the image in a particular basis (perhaps a wavelet basis), then keep the most important components and throw away the less important ones. Lbc = c d

Graph Laplacians Laplacian of graph G LG = §e2E Le = -1 for every edge b Laplacian of graph G a b c d a 2 -1 3 1 -1 for every edge b LG = §e2E Le = Let me begin the talk by motivating what sparsification is and why it’s important. Sparsification is a concept that’s familiar from our daily lives: you want to take some object that’s dense, and replace by an object that’s sparser, but almost as good as the original object. Here’s an example I learned about when I bought my first house a few years ago: floor joists used to be made of solid wood, but now they’re “engineered”, often with a truss structure, often with other materials. Another familiar example is image compression: we express the image in a particular basis (perhaps a wavelet basis), then keep the most important components and throw away the less important ones. c degree of node d

Spectrally-thin trees 5 -1 4 6 7 6 -1 -5 5 -3 2 8 -8 1 10 A spanning tree T is ®-spectrally-thin if LT ¹ ®¢LG Effective Resistance from s to t: Rst = voltage difference when a 1-amp current source placed between s and t Theorem [Harvey-Olver '14]: Every n-vertex graph has an ®-spectrally-thin spanning tree with ®= . Uses matrix concentration bounds. Algorithmic. Let me begin the talk by motivating what sparsification is and why it’s important. Sparsification is a concept that’s familiar from our daily lives: you want to take some object that’s dense, and replace by an object that’s sparser, but almost as good as the original object. Here’s an example I learned about when I bought my first house a few years ago: floor joists used to be made of solid wood, but now they’re “engineered”, often with a truss structure, often with other materials. Another familiar example is image compression: we express the image in a particular basis (perhaps a wavelet basis), then keep the most important components and throw away the less important ones.

Spectrally-thin trees 5 -1 4 6 7 6 -1 -5 5 -3 2 8 -8 1 10 A spanning tree T is ®-spectrally-thin if LT ¹ ®¢LG Effective Resistance from s to t: Rst = voltage difference when a 1-amp current source placed between s and t Theorem: Every n-vertex graph has an ®-spectrally-thin spanning tree with ®= . Follows from Kadison-Singer solution of MSS'13. Not algorithmic. Let me begin the talk by motivating what sparsification is and why it’s important. Sparsification is a concept that’s familiar from our daily lives: you want to take some object that’s dense, and replace by an object that’s sparser, but almost as good as the original object. Here’s an example I learned about when I bought my first house a few years ago: floor joists used to be made of solid wood, but now they’re “engineered”, often with a truss structure, often with other materials. Another familiar example is image compression: we express the image in a particular basis (perhaps a wavelet basis), then keep the most important components and throw away the less important ones.

Asymmetric Traveling Salesman Problem Recent breakthrough: [Ansari, Oveis-Gharan Dec 2014] Show how to build on the O(1)-spectrally-thin tree result to approximate optimal weight of an ATSP solution to within poly(log log n) of optimal. But, no algorithm to find the actual sequence of vertices!

Our Main Result Let P½[0,1]m be a matroid base polytope (e.g., convex hull of characteristic vectors of spanning trees) Let A1,…, Am be PSD matrices of size nxn. Define and Q;. There is an extreme point Â(S) of P with

Our Main Result Let P½[0,1]m be a matroid base polytope. Let A1,…, Am be PSD matrices of size nxn. Define and Q;. There is an extreme point Â(S) of P with What is dependence on ®? Easy: ® ¸ 1.5, even with n=2. Standard random matrix theory: ® = O(log n). Our result: Ideally: ®<2. This would solve Kadison-Singer problem. MSS ‘13: Solved Kadison-Singer, achieve ® = O(1).

Our Main Result Let P½[0,1]m be a matroid base polytope. Let A1,…, Am be PSD matrices of size nxn. Define and Q;. There is an extreme point Â(S) of P with , Furthermore, there is a random process that starts at any x02Q and terminates after m steps at such a point Â(S), whp. each step of this process can be performed algorithmically. the entire process can be derandomized.

Pipage rounding Let P be any matroid polytope. Given fractional x [Ageev-Svirideno ‘04, Srinivasan ‘01, Calinescu et al. ‘07, Chekuri et al. ‘09] Let P be any matroid polytope. Given fractional x Find coordinates a and b s.t. line z  x + z ( ea – eb ) stays in current face Find two points where line leaves P Randomly choose one of those points s.t. expectation is x Repeat until x = ÂT is integral x is a martingale: expectation of final ÂT is original fractional x. ÂT1 ÂT6 ÂT2 x ÂT3 ÂT5 ÂT4

Pessimistic estimators Definition: “Pessimistic Estimator” Let E µ {0,1}m be an event. Let D(x) be the product distribution on {0,1}m with expectation x. Then g : [0,1]m ! R is a pessimistic estimator for E if Example: If E is the event { x : wT x>t } then Chernoff bounds give the pessimistic estimator

Concavity under swaps Definition: A functionf : Rm ! R is concave under swaps if z ! f( x + z(ea-eb) ) is concave 8x2P, 8a, b2[m]. Example: is concave under swaps. Pipage Rounding: Let X0 be initial point and ÂT be final point visited by pipage rounding. Claim: If f concave under swaps then E[f(ÂT)] · f(X0). [by Jensen] Pessimistic Estimators: Let E be an event and g a pessimistic estimator for E. Claim: Suppose g is concave under swaps. Then Pr[ ÂT 2 E ] · g(X0).

Matrix Pessimistic Estimators Special case of Tropp ‘12: Let A1,…,Am be nxn PSD matrices. Let D(x) be the product distribution on {0,1}m with expectation x. Let Suppose ¹¢Ai ¹ M. Let Then and . Pessimistic estimator Main Technical Result: gt,µ is concave under swaps. ) Tropp’s bound for independent sampling also achieved by pipage rounding

Our Variant of Lieb’s Theorem: PD Our Variant of Lieb’s Theorem:

Questions Does Tropp’s matrix concentration bound hold in a negatively dependent scenario? Does our variant of Lieb’s theorem have other uses? O(maxe Re)-spectrally thin trees exist by MSS’13. Can they be constructed algorithmically?