Discrepancy and Optimization

Slides:



Advertisements
Similar presentations
Discrepancy and SDPs Nikhil Bansal (TU Eindhoven, Netherlands ) August 24, ISMP 2012, Berlin.
Advertisements

Enumerative Lattice Algorithms in any Norm via M-Ellipsoid Coverings Daniel Dadush (CWI) Joint with Chris Peikert and Santosh Vempala.
Lecture 24 Coping with NPC and Unsolvable problems. When a problem is unsolvable, that's generally very bad news: it means there is no general algorithm.
1/26 Constructive Algorithms for Discrepancy Minimization Nikhil Bansal (IBM)
1/17 Deterministic Discrepancy Minimization Nikhil Bansal (TU Eindhoven) Joel Spencer (NYU)
Discrepancy Minimization by Walking on the Edges Raghu Meka (IAS/DIMACS) Shachar Lovett (IAS)
Approximation Algorithms for Capacitated Set Cover Ravishankar Krishnaswamy (joint work with Nikhil Bansal and Barna Saha)
N. Bansal 1, M. Charikar 2, R. Krishnaswamy 2, S. Li 3 1 TU Eindhoven 2 Princeton University 3 TTIC Midwest Theory Day, Purdue, May 3, 2014.
Introduction to Approximation Algorithms Lecture 12: Mar 1.
Semidefinite Programming
Recent Development on Elimination Ordering Group 1.
1 Optimization problems such as MAXSAT, MIN NODE COVER, MAX INDEPENDENT SET, MAX CLIQUE, MIN SET COVER, TSP, KNAPSACK, BINPACKING do not have a polynomial.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2002 Review Lecture Tuesday, 12/10/02.
High Performance Computing 1 Parallelization Strategies and Load Balancing Some material borrowed from lectures of J. Demmel, UC Berkeley.
Randomness in Computation and Communication Part 1: Randomized algorithms Lap Chi Lau CSE CUHK.
(work appeared in SODA 10’) Yuk Hei Chan (Tom)
Experts and Boosting Algorithms. Experts: Motivation Given a set of experts –No prior information –No consistent behavior –Goal: Predict as the best expert.
Minimizing Flow Time on Multiple Machines Nikhil Bansal IBM Research, T.J. Watson.
A tale of 2-dimensional (Guillotine) bin packing Nikhil Bansal (IBM) Andrea Lodi (Univ. of Bologna, Italy) Maxim Sviridenko (IBM)
Daniel Kroening and Ofer Strichman Decision Procedures An Algorithmic Point of View Deciding ILPs with Branch & Bound ILP References: ‘Integer Programming’
Discrepancy Minimization by Walking on the Edges Raghu Meka (IAS & DIMACS) Shachar Lovett (IAS)
1/30 Discrepancy and SDPs Nikhil Bansal (TU Eindhoven)
Beating the Union Bound by Geometric Techniques Raghu Meka (IAS & DIMACS)
1 The Santa Claus Problem (Maximizing the minimum load on unrelated machines) Nikhil Bansal (IBM) Maxim Sviridenko (IBM)
C&O 355 Mathematical Programming Fall 2010 Lecture 17 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
Decision Procedures An Algorithmic Point of View
C&O 355 Mathematical Programming Fall 2010 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A.
Correlation testing for affine invariant properties on Shachar Lovett Institute for Advanced Study Joint with Hamed Hatami (McGill)
Stochastic Algorithms Some of the fastest known algorithms for certain tasks rely on chance Stochastic/Randomized Algorithms Two common variations – Monte.
Approximation schemes Bin packing problem. Bin Packing problem Given n items with sizes a 1,…,a n  (0,1]. Find a packing in unit-sized bins that minimizes.
1 Introduction to Approximation Algorithms. 2 NP-completeness Do your best then.
Packing Rectangles into Bins Nikhil Bansal (CMU) Joint with Maxim Sviridenko (IBM)
CSE 589 Part VI. Reading Skiena, Sections 5.5 and 6.8 CLR, chapter 37.
Implicit Hitting Set Problems Richard M. Karp Erick Moreno Centeno DIMACS 20 th Anniversary.
1/30 Constructive Algorithms for Discrepancy Minimization Nikhil Bansal (IBM)
FATCOP: A Mixed Integer Program Solver Michael FerrisQun Chen University of Wisconsin-Madison Jeffrey Linderoth Argonne National Laboratories.
geometric representations of graphs
TU/e Algorithms (2IL15) – Lecture 12 1 Linear Programming.
Linear Programming Piyush Kumar Welcome to CIS5930.
TU/e Algorithms (2IL15) – Lecture 12 1 Linear Programming.
Approximation algorithms
Optimal Acceleration and Braking Sequences for Vehicles in the Presence of Moving Obstacles Jeff Johnson, Kris Hauser School of Informatics and Computing.
Lecture 1 (Part 1) Introduction/Overview Tuesday, 9/9/08
Introduction to Approximation Algorithms
Lap Chi Lau we will only use slides 4 to 19
Integer Programming An integer linear program (ILP) is defined exactly as a linear program except that values of variables in a feasible solution have.
Abolfazl Asudeh Azade Nazi Nan Zhang Gautam DaS
Optimization problems such as
Topics in Algorithms Lap Chi Lau.
Introduction to Machine Learning
Anticoloring Problems on Graphs
Algorithms for hard problems
Generalization and adaptivity in stochastic convex optimization
Distributed Submodular Maximization in Massive Datasets
Design and Analysis of Computer Algorithm (CS575-01)
Background: Lattices and the Learning-with-Errors problem
Randomized Algorithms CS648
Nikhil Bansal, Shashwat Garg, Jesper Nederlof, Nikhil Vyas
Size Limited Bounding Polygons for Planar Point Sets
Chapter 6. Large Scale Optimization
Probabilistic Models with Latent Variables
Integer Programming (정수계획법)
Instructors: Fei Fang (This Lecture) and Dave Touretzky
geometric representations of graphs
Algorithm design (computational geometry)
Integer Programming (정수계획법)
On Approximating Covering Integer Programs
Approximation Algorithms
Integer Programming (IP)
Discrete Optimization
Presentation transcript:

Discrepancy and Optimization Nikhil Bansal (CWI and TU/e)

Discrepancy: What is it? Study of irregularities in approximating the continuous by the discrete. Original motivation: Numerical Integration/ Sampling How well can you approximate a region by discrete points ?

Discrepancy: What is it? Problem: How uniformly can you distribute n points in a grid. “Uniform” : For every axis-parallel rectangle R | (# points in R) - (Area of R) | should be low. Discrepancy: Max over rectangles R |(# points in R) – (Area of R)| R n1/2 n1/2

Distributing points in a grid Problem: How uniformly can you distribute n points in a grid. “Uniform” : For every axis-parallel rectangle R | (# points in R) - (Area of R) | should be low. n= 64 points Uniform Random Van der Corput Set n1/2 discrepancy n1/2 (loglog n)1/2 O(log n) discrepancy!

Quasi-Monte Carlo Methods n random samples (Monte Carlo) : Error ∝ 1 𝑛 Quasi-Monte Carlo Methods* : ∝ 𝐷𝑖𝑠𝑐 𝑛 Extensive research area. *Different constant of proportionality

Combinatorial Discrepancy Universe: U= [1,…,n] Subsets: S1,S2,…,Sm Color elements red/blue so each set is colored as evenly as possible. Given : [n] ! { -1,+1} Disc (𝜒) = maxS |i2S (i)| = maxS  𝑆 Disc (set system) = min𝜒 maxS  𝑆 S4 S1 S2

Tusnady’s problem Input: n points placed arbitrarily in a grid. Sets = axis-parallel rectangles Discrepancy: max over rect. R ( |# red in R - # blue in R| ) Random gives about O(n1/2 log1/2 n) Very long line of work O(lo g 4 n) [Beck 80’s] ... O(log2.5 n) [Matousek’99] O( log 2 n ) [B., Garg’16] O(log1.5 n) [Nikolov’17]

Application: Dynamic Data Structures N weighted points in a 2-d region. Weights updated over time. Query: Given an axis-parallel rectangle R, determine the total weight on points in R. Goal: Preprocess (in a data structure) Low query time Low update time (upon weight change)

Example Line: Interval queries Trivial: Query Time= O(n) Update Time = 1 Query = 1 Update = O( 𝑛 2 ) (Table of entries W[a,b] ) Query = 2 Update = O(n) (W[a,b] = W[0,b] – W[0,a]) Query = O(log n) Update = O(log n) Recursively for 2-d. 𝑂 log 2 𝑛 , log 2 𝑛

What about other queries? Circles arbitrary rectangles aligned triangle Turns out 𝑡 𝑞 𝑡 𝑢 ≥ 𝑛 1/2 log 2 𝑛 Reason: Set system S formed by query sets & points has large discrepancy (about 𝑛 1/4 ) Larsen-Green’11: 𝑡 𝑞 𝑡 𝑢 ≥ 𝑑𝑖𝑠𝑐 𝑆 2 log 2 𝑛

Applications CS: Computational Geometry, Approximation, Complexity, Differential Privacy, Pseudo-Randomness, … Math: Combinatorics, Finance, Dynamical Systems, Number Theory, Ramsey Theory, Algebra, Measure Theory, …

Matrix View Given any matrix A, Rows: sets Columns: elements Given any matrix A, find coloring 𝑥 ∈ −1,1 𝑛 , to minimize 𝐴𝑥 ∞

Optimization Jobs Machines Travelling Salesman Scheduling Clustering NP-Hard Min cx Ax = b x ∈ {0,1} Polynomial time Min cx Ax = b x ∈ [0,1] relaxation rounding

Rounding Lovasz-Spencer-Vesztermgombi’86: Given any matrix A, and 𝑥∈ 𝑅 𝑛 , can round x to 𝑥 ∈ 𝑍 𝑛 s.t. 𝐴𝑥 –𝐴 𝑥 ∞ < Herdisc 𝐴 Proof: Round the bits of x one by one. 𝑥 1 : blah .0101101 𝑥 2 : blah .1101010 … 𝑥 𝑛 : blah .0111101 Error = herdisc(A) ( 1 2 𝑘 + 1 2 𝑘−1 + … + 1 2 ) 𝑥 Ax=b x (-1) Key Point: Low discrepancy coloring guides our updates! A (+1)

Rounding: The Catch LSV’86 result guarantees existence of good rounding. But, how to find it efficiently (poly time)? Nothing known until recently.

Questions around Discrepancy bounds Combinatorial: Show good coloring exists Algorithmic: Find coloring in poly time Lower bounds on discrepancy (do not focus here)

Combinatorial (3 generations) 0) Linear Algebra (Iterated Rounding) [Steinitz, Beck-Fiala, Barany, …] 1) Partial Coloring Method: Beck/Spencer early 80’s: Probabilistic Method + Pigeonhole Gluskin’87: Convex Geometric Approach Very versatile (black-box) Loss adds over O(log n) iterations 2) Banaszczyk’98: Based on a deep convex geometric result Produces full coloring directly (also black-box)

Brief History (combinatorial) Method Tusnady (rectangles) Beck-Fiala (low deg. system) Linear Algebra log 4 𝑛 t Partial Coloring log 2.5 𝑛 [Matousek’99] t1/2 log n Banaszczyk log 1.5 𝑛 [Nikolov’17] (t log n)1/2 [Banaszczyk’98] Lower bound log 𝑛 t1/2

Algorithmic history Linear Algebraic Methods: Already algorithmic Partial Coloring now constructive Bansal’10: SDP + Random walk Lovett Meka’12: Random walk + linear algebra Rothvoss’14: Convex geometric Many others by now [Harvey, Schwartz, Singh], [Eldan, Singh] Banaszczyk based approaches: [B.-Dadush-Garg’16]: 𝑡 log 𝑛 1/2 algorithm for (specific) Beck-Fiala problem [B.-Garg’17]: A more general framework, e.g. 𝑂 log 2 𝑛 for Tusnady. [B. Dadush Garg Nikolov’ 18]: Algorithm for general Banaszczyk.

Several Successes Meta-Problem: Ideas for discrepancy lead to new rounding techniques for optimization problems Bin Packing Combines randomized and iterated rounding [B., Nagarajan’16, …] (two very powerful but quite orthogonal approaches) Big Goal: Currently algorithm design: quite adhoc, problem specific approaches. Discover general principles that unify + improve existing approaches 4 3 2 4 ! 1 2 3 1

Optimization in Practice Model your program as (mixed) integer program. Use solver like CPLEX or Gurobi Various heuristics: Branch & bound, Cutting planes, Preprocessing, … Big mystery why they work so well (on real world instances) Could we explain/understand this success theoretically? Could techniques from discrepancy help in improving heuristics? Workshop: Discrepancy theory and Integer Prog. (Jun 11-15, CWI) (more info on Daniel Dadush’s webpage)

Questions!