Download presentation

Presentation is loading. Please wait.

Published byKeenan Buckland Modified over 3 years ago

1
Discrepancy and SDPs Nikhil Bansal (TU Eindhoven, Netherlands ) August 24, ISMP 2012, Berlin

2
Outline Discrepancy Theory What is it Applications Basic Results (non-constructive) SDP connection Algorithms Lower Bounds 2/40

3
Discrepancy Theory: What is it? Study of discrepancy between self-perception and reality 3/40

4
Discrepancy: What is it? Study of irregularities in approximating the continuous by the discrete. Historical motivation: Numerical Integration/ Sampling How well can you approximate a region by discrete points ? 4/40

5
Discrepancy: What is it? Problem: How uniformly can you distribute points in a grid. “Uniform” : For every axis-parallel rectangle R | (# points in R) - (Area of R) | should be low. n 1/2 R Discrepancy: Max over rectangles R |(# points in R) – (Area of R)| 5/40

6
Distributing points in a grid Problem: How uniformly can you distribute points in a grid. “Uniform” : For every axis-parallel rectangle R | (# points in R) - (Area of R) | should be low. Uniform RandomVan der Corput Set n= 64 points n 1/2 discrepancyn 1/2 (loglog n) 1/2 O(log n) discrepancy! 6/40

7
Quasi-Monte Carlo Methods *Different constant of proportionality 7/40

8
Discrepancy: Example 2 Input: n points placed arbitrarily in a grid. Color them red/blue such that each axis-parallel rectangle is colored as evenly as possible Discrepancy: max over rect. R ( | # red in R - # blue in R | ) Continuous: Color each element 1/2 red and 1/2 blue (0 discrepancy) Discrete: Random has about O(n 1/2 log 1/2 n) Can achieve O(log 2.5 n) Why do we care? 8/40

9
Combinatorial Discrepancy S1S1 S2S2 S3S3 S4S4 9/40

10
Combinatorial Discrepancy Set system: A = {0,1} incidence matrix 10/40

11
Applications CS: Computational Geometry, Comb. Optimization, Monte-Carlo simulation, Machine learning, Complexity, Pseudo-Randomness, … Math: Dynamical Systems, Combinatorics, Mathematical Finance, Number Theory, Ramsey Theory, Algebra, Measure Theory, … 11/40

12
Hereditary Discrepancy A1A2…A1A2… 1 2 … n A’ 1 A’ 2 … 1’ 2’ … n’ But not so robust 12/40 Discrepancy = 0

13
Two Applications 13/40

14
Rounding Ax=b A 14/40 (-1) (+1) Key Point: Low discrepancy coloring guides our updates! x

15
Rounding 15/40

16
Discrepancy and optimization 16/40

17
Dynamic Data Structures N weighted points in a 2-d region. Weights updated over time. Query: Given an axis-parallel rectangle R, determine the total weight on points in R. Goal: Preprocess (in a data structure) 1)Low query time 2)Low update time (upon weight change) 17/40

18
Example 18/40

19
What about other queries? 19/40

20
Bounding Discrepancy 20/40

21
General set system 21/40

22
(Previous) Best Algorithm 22/40

23
Better Colorings Exist! [Spencer 85]: (Six standard deviations suffice) Any system with n sets has discrepancy · 6n 1/2 (In general for arbitrary m, discrepancy = O(n 1/2 log(m/n) 1/2 ) Tight: For m=n, cannot beat 0.5 n 1/2 (Hadamard Matrix) Inherently non-constructive proof (counting) Powerful Entropy Method. Question: Can we find it algorithmically ? Certain algorithms do not work [Spencer] Conjecture [Alon-Spencer]: May not be possible. 23/40 Space of colorings

24
Results General Technique: k-permutation problem [Spencer, Srinivasan,Tetali] geometric problems, Beck Fiala setting (Srinivasan’s bound) … 24/40

25
SDPs 25/40

26
Relaxations: LPs and SDPs Yet, SDPs will be a major tool. 26/40

27
Punch line 27/40

28
Algorithm (at high level) Cube: {-1,+1} n Analysis: Few steps to reach a vertex (walk has high variance) Disc( S i ) does a random walk (with low variance) start finish Algorithm: “Sticky” random walk Each step generated by rounding a suitable SDP Move in various dimensions correlated, e.g. t 1 + t 2 ¼ 0 Each dimension: An Element Each vertex: A Coloring 28/40

29
An SDP Hereditary disc. ) the following SDP is feasible Obtain v i 2 R n 29/40

30
Idea Which vector g to project on? Lemma: If g 2 R n is a random Gaussian, for any v 2 R n, g ¢ v is distributed as N(0, |v| 2 ). Pf: N(0,a 2 ) + N(0,b 2 ) = N(0,a 2 +b 2 ) g ¢ v = i v(i) g i » N(0, i v(i) 2 ) 30/40

31
Properties of Rounding Lemma: If g 2 R n is a random Gaussian, for any v 2 R n, g ¢ v is distributed as N(0, |v| 2 ) 1.Each i » N(0, ) 2.For each set S, i 2 S i = g ¢ ( i 2 S v i ) » N(0, · 2 ) (std deviation · ) SDP: |v i | 2 = 1 | i 2 S v i | 2 · 2 Recall: i = g ¢ v i ’s will guide our updates to x. 31/40

32
Algorithm Overview Construct coloring iteratively. Initially: Start with coloring x 0 = (0,0,0, …,0) at t = 0. At Time t: Update coloring as x t = x t-1 + ( t 1,…, t n ) ( tiny: 1/n suffices) x(i) x t (i) = ( 1 i + 2 i + … + t i ) Color of element i: Does random walk over time with step size ¼ Fixed if reaches -1 or +1. time +1 Set S: x t (S) = i 2 S x t (i) does a random walk w/ step N(0, · 2 ) 32/40

33
Analysis Consider time T = O(1/ 2 ) Claim 1: With prob. ½, an element reaches -1 or +1. Pf: Each element doing random walk (martingale) with size ¼ Recall: Random walk with step 1, is ¼ t 1/2 away in t steps. Claim 2: Each set has O( ) discrepancy in expectation. Pf: For each S, x t (S) doing random walk with step size ¼ 33/40

34
Recap At each step of walk, formulate SDP on floating variables. SDP solution -> Guides the walk Properties of walk: High Variance -> Quick convergence Low variance for discrepancy on sets -> Low discrepancy 34/40 start finish

35
Refinements Spencer’s six std deviations result: Recall: Want O(n 1/2 ) discrepancy, but random coloring gives n 1/2 (log n) 1/2 Previous approach seems useless: Expected discrepancy for a set O(n 1/2 ), but some random walks will deviate by up to (log n) 1/2 factor Tune down the variance of dangerous sets (not too many) Entropy Method -> SDP still feasible. 0 20n 1/2 30n 1/2 35n 1/2 … Danger 1 Danger 2 Danger 3 …

36
Further Developments Can be derandomized [Bansal-Spencer’11] Our algorithm still uses the Entrpoy method. Gives no new proof of Spencer’s result. Is there a purely constructive proof ? Lovett Meka’12: Yes. Gaussian random walks + Linear Algebra 36/40

37
Matousek Lower Bound 37/40

38
Detlb 38/40

39
In Conclusion 39/40

40
Thank you for your attention 40/40

Similar presentations

OK

Milan Vojnović Microsoft Research Cambridge Collaborators: E. Perron and D. Vasudevan 1 Consensus – with Limited Processing and Signalling.

Milan Vojnović Microsoft Research Cambridge Collaborators: E. Perron and D. Vasudevan 1 Consensus – with Limited Processing and Signalling.

© 2018 SlidePlayer.com Inc.

All rights reserved.

To make this website work, we log user data and share it with processors. To use this website, you must agree to our Privacy Policy, including cookie policy.

Ads by Google

Ppt on disaster management for class 10 download Ppt on kinetic energy for class 9 Ppt on emerging technologies in computer science Ppt on social contract theory rawls Ppt on sea level rise graph Ppt on conventional energy sources Ppt on save environment photos Ppt on indian politics swami Animal sounds for kids ppt on batteries Raster scan and random scan display ppt on tv