Warren Center for Network and Data Sciences

Slides:



Advertisements
Similar presentations
Hardness of Reconstructing Multivariate Polynomials. Parikshit Gopalan U. Washington Parikshit Gopalan U. Washington Subhash Khot NYU/Gatech Rishi Saket.
Advertisements

Routing Complexity of Faulty Networks Omer Angel Itai Benjamini Eran Ofek Udi Wieder The Weizmann Institute of Science.
Hardness of testing 3- colorability in bounded degree graphs Andrej Bogdanov Kenji Obata Luca Trevisan.
1 Transitive-Closure Spanners Sofya Raskhodnikova Penn State University Joint work with Arnab Bhattacharyya MIT Elena Grigorescu MIT Kyomin Jung MIT David.
Property Testing of Data Dimensionality Robert Krauthgamer ICSI and UC Berkeley Joint work with Ori Sasson (Hebrew U.)
Property Testing and Communication Complexity Grigory Yaroslavtsev
Lower Bounds for Testing Properties of Functions on Hypergrids Grigory Yaroslavtsev Joint with: Eric Blais (MIT) Sofya Raskhodnikova.
Grigory Yaroslavtsev Joint work with Piotr Berman and Sofya Raskhodnikova.
Distributional Property Estimation Past, Present, and Future Gregory Valiant (Joint work w. Paul Valiant)
Property Testing on Product Distributions: Optimal Testers for Bounded Derivative Properties Deeparnab Chakrabarty Microsoft Research Bangalore Kashyap.
Christian Sohler | Every Property of Hyperfinite Graphs is Testable Ilan Newman and Christian Sohler.
New Algorithms and Lower Bounds for Monotonicity Testing of Boolean Functions Rocco Servedio Joint work with Xi Chen and Li-Yang Tan Columbia University.
Learning and Fourier Analysis Grigory Yaroslavtsev CIS 625: Computational Learning Theory.
Testing the Diameter of Graphs Michal Parnas Dana Ron.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 8 May 4, 2005
Testing of Clustering Noga Alon, Seannie Dar Michal Parnas, Dana Ron.
Sublinear time algorithms Ronitt Rubinfeld Blavatnik School of Computer Science Tel Aviv University TexPoint fonts used in EMF. Read the TexPoint manual.
Michael Bender - SUNY Stony Brook Dana Ron - Tel Aviv University Testing Acyclicity of Directed Graphs in Sublinear Time.
Testing Metric Properties Michal Parnas and Dana Ron.
On Proximity Oblivious Testing Oded Goldreich - Weizmann Institute of Science Dana Ron – Tel Aviv University.
On Testing Convexity and Submodularity Michal Parnas Dana Ron Ronitt Rubinfeld.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint work with Mira Gonen Dana Ron Tel-Aviv University.
1 Algorithmic Aspects in Property Testing of Dense Graphs Oded Goldreich – Weizmann Institute Dana Ron - Tel-Aviv University.
1 On the Benefits of Adaptivity in Property Testing of Dense Graphs Joint works with Mira Gonen and Oded Goldreich Dana Ron Tel-Aviv University.
Lower Bounds for Property Testing Luca Trevisan U C Berkeley.
Christian Sohler 1 University of Dortmund Testing Expansion in Bounded Degree Graphs Christian Sohler University of Dortmund (joint work with Artur Czumaj,
Some 3CNF Properties are Hard to Test Eli Ben-Sasson Harvard & MIT Prahladh Harsha MIT Sofya Raskhodnikova MIT.
Dana Moshkovitz, MIT Joint work with Subhash Khot, NYU.
Approximating the MST Weight in Sublinear Time Bernard Chazelle (Princeton) Ronitt Rubinfeld (NEC) Luca Trevisan (U.C. Berkeley)
Complexity and Efficient Algorithms Group / Department of Computer Science Approximating Structural Properties of Graphs by Random Walks Christian Sohler.
Transitive-Closure Spanner of Directed Graphs Kyomin Jung KAIST 2009 Combinatorics Workshop Joint work with Arnab Bhattacharyya MIT Elena Grigorescu MIT.
1 Sublinear Algorithms Lecture 1 Sofya Raskhodnikova Penn State University TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
1/19 Minimizing weighted completion time with precedence constraints Nikhil Bansal (IBM) Subhash Khot (NYU)
Artur Czumaj DIMAP DIMAP (Centre for Discrete Maths and it Applications) Computer Science & Department of Computer Science University of Warwick Testing.
狄彥吾 (Yen-Wu Ti) 華夏技術學院資訊工程系 Property Testing on Combinatorial Objects.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Complexity and Efficient Algorithms Group / Department of Computer Science Testing the Cluster Structure of Graphs Christian Sohler joint work with Artur.
Algorithms for Big Data: Streaming and Sublinear Time Algorithms
On Sample Based Testers
Property Testing (a.k.a. Sublinear Algorithms )
New Characterizations in Turnstile Streams with Applications
Dana Ron Tel Aviv University
On Testing Dynamic Environments
Vitaly Feldman and Jan Vondrâk IBM Research - Almaden
Approximating the MST Weight in Sublinear Time
Lecture 22: Linearity Testing Sparse Fourier Transform
Finding Cycles and Trees in Sublinear Time
Understanding Generalization in Adaptive Data Analysis
On Approximating the Number of Relevant Variables in a Function
From dense to sparse and back again: On testing graph properties (and some properties of Oded)
Lecture 18: Uniformity Testing Monotonicity Testing
Lecture 10: Sketching S3: Nearest Neighbor Search
CIS 700: “algorithms for Big Data”
CIS 700: “algorithms for Big Data”
Linear sketching with parities
CSCI B609: “Foundations of Data Science”
Locally Decodable Codes from Lifting
CSCI B609: “Foundations of Data Science”
Sublinear Algorihms for Big Data
Linear sketching over
CSCI B609: “Foundations of Data Science”
Linear sketching with parities
CSCI B609: “Foundations of Data Science”
The Subgraph Testing Model
Every set in P is strongly testable under a suitable encoding
Lecture 6: Counting triangles Dynamic graphs & sampling
Generalization bounds for uniformly stable algorithms
Clustering.
Sublinear Algorihms for Big Data
Presentation transcript:

Warren Center for Network and Data Sciences 𝑳 𝒑 -Testing With P. Berman and S. Raskhodnikova (STOC’14+). Grigory Yaroslavtsev Warren Center for Network and Data Sciences http://grigory.us Plug in my job search and Warren Center.

Property Testing [Goldreich, Goldwasser, Ron; Rubinfeld, Sudan] NO YES Randomized Algorithm Accept with probability ≥ 𝟐 𝟑 Reject with probability ≥ 𝟐 𝟑 ⇒ ⇒ YES NO Property Tester 𝝐-close Accept with probability ≥ 𝟐 𝟑 Reject with probability ≥ 𝟐 𝟑 Don’t care A randomized algorithm distinguishes with some constant probability (e.g. 2/3) the two scenarios: monotone vs. not monotone. The problem with a randomized algorithm is that in the worst case it has too look in the entire data, which is too big. So the running time of such algorithm would be at least linear. In order to be able do design sublinear time algorithms one has to relax its definition. The first such relaxation, called property testing, was introduced by bla-bla-bla. A parameter \epsilon allows to specify approximation in the decision problem. If the data is monotone a property testing algorithm still accepts. 𝝐-close : ≤𝝐 fraction has to be changed to become YES

Which stocks were growing? Given time-series data we can ask ourselves a question: does it follow a monotone trend? Let’s look at the dynamics of stocks and see whether they have demonstrated robust growth. We are looking at the data from Google Finance over the 5 years following recession. Apple was growing steadily, but then Data from http://finance.google.com

Property testing: testing monotonicity? Given time-series data we can ask ourselves a question: does it follow a monotone trend? Let’s look at the dynamics of stocks and see whether they have demonstrated robust growth. We are looking at the data from Google Finance over the 5 years following recession. Apple was growing steadily, but then Data from http://finance.google.com

Tolerant Property Testing [Parnas, Ron, Rubinfeld] ⇒ YES NO Property Tester 𝝐-close Accept with probability ≥ 𝟐 𝟑 Reject with probability ≥ 𝟐 𝟑 Don’t care ⇒ Tolerant Property Tester Accept with probability ≥ 𝟐 𝟑 Reject with probability ≥ 𝟐 𝟑 Don’t care NO 𝝐 𝟏 -close ( 𝝐 𝟏 , 𝝐 𝟐 )-close YES 𝝐-close : ≤𝝐 fraction has to be changed to become YES

Tolerant monotonicity testing? Data from http://finance.google.com

Tolerant “ 𝑳 𝟏 Property Testing” Tolerant “ 𝑳 𝟏 Property Tester” 𝒇:{1, …, 𝑛}→ 0,1 𝑷 = class of monotone functions 𝑑𝑖𝑠 𝑡 1 𝒇,𝑷 = min 𝒈∈𝑷 𝒇 −𝒈 1 𝑛 𝝐-close: 𝑑𝑖𝑠 𝑡 1 𝒇,𝑷 ≤𝝐 More general: distance approximation Even more general: isotonic regression Accept with probability ≥ 𝟐 𝟑 Reject with probability ≥ 𝟐 𝟑 ⇒ Don’t care YES 𝝐 𝟏 -close ⇒ ( 𝝐 𝟏 , 𝝐 𝟐 )-close NO

𝐿 1 -Isotonic Regression Pool Adjacent Violators Algorithm Running time O 𝑛 log 𝑛 [Folklore] Available in Matlab/R packages

New 𝐿 𝑝 -Testing Model for Real-Valued Data Generalizes standard Hamming testing For 𝑝>0 still has a probabilistic interpretation: 𝑑 𝑝 𝑓,𝑔 = 𝐄 𝒇−𝒈 𝒑 1/𝑝 Compatible with existing PAC-style learning models that have 𝐿 𝑝 -error (preprocessing for model selection) For Boolean functions, 𝑑 0 𝑓,𝑔 = 𝑑 𝑝 𝑓,𝑔 𝑝 . Various distances used widely in distribution testing

Our Contributions Relationships between 𝐿 𝒑 -testing models Algorithms 𝐿 𝒑 -testers for 𝒑≥1 monotonicity, Lipschitzness, convexity Tolerant 𝐿 𝒑 -tester for 𝒑≥1 monotonicity in 1D (sublinear algorithm for isotonic regression) monotonicity in 2D Our 𝐿 𝒑 -testers beat lower bounds for Hamming testers Simple algorithms backed up by involved analysis Uniformly sampled (or easy to sample) data suffices Nearly tight lower bounds in many cases

Implications for Hamming Testing Some techniques/results carry over to Hamming testing Improvement on Levin’s work investment strategy Connectivity of bounded-degree graphs [Goldreich, Ron ‘02] Properties of images [Raskhodnikova ‘03] Multiple-input problems [Goldreich ‘13] First example of monotonicity testing problem where adaptivity helps Improvements to Hamming testers for Boolean functions

Definitions 𝒇: 𝐷→ 0,1 (D = finite domain/poset) 𝒇 𝒑 =( 𝑥∈ 𝐷 𝒇 𝑥 𝒑 ) 1/𝒑 , for 𝒑≥1 𝒇 𝟎 = Hamming weight (# of non-zero values) Property 𝑷 = class of functions (monotone, convex, Lipschitz, …) 𝑑𝑖𝑠 𝑡 𝒑 𝒇,𝑷 = min 𝑔∈𝑷 ||𝒇 −𝑔|| 𝒑 1 𝒑

Relationships: 𝐿 𝑝 -Testing 𝑄 𝒑 (𝑷,𝝐) = query complexity of 𝐿 𝒑 -testing property 𝑷 at distance 𝝐 𝑄 𝟏 (𝑷,𝝐)≤ 𝑄 𝟎 (𝑷,𝝐) 𝑄 𝟏 (𝑷,𝝐)≤ 𝑄 𝟐 (𝑷,𝝐) (Cauchy-Shwarz) 𝑄 𝟏 (𝑷,𝝐)≥ 𝑄 𝟐 (𝑷, 𝝐 ) Boolean functions 𝒇:𝐷→ 0,1 𝑄 𝟎 (𝑷,𝝐) = 𝑄 𝟏 (𝑷,𝝐) = 𝑄 𝟐 (𝑷, 𝝐 )

Relationships: Tolerant 𝐿 𝑝 -Testing 𝑄 𝒑 (𝑷, 𝝐 𝟏 , 𝝐 𝟐 ) = query complexity of tolerant 𝐿 𝒑 -testing property 𝑷 with distance parameters 𝝐 𝟏 , 𝝐 𝟐 No general relationship between tolerant 𝐿 𝟏 -testing and tolerant Hamming testing 𝐿 𝒑 -testing for 𝒑>1 is close in complexity to 𝐿 𝟏 -testing 𝑄 𝟏 (𝑷, 𝜺 𝟏 𝒑 , 𝜺 𝟐 )≤𝑄 𝒑 (𝑷, 𝜺 𝟏 , 𝜺 𝟐 )≤ 𝑄 𝟏 (𝑷, 𝜺 𝟏 , 𝜺 𝟐 𝒑 ) For Boolean functions 𝒇:𝐷→ 0,1 𝑄 𝟎 (𝑷, 𝜺 𝟏 , 𝜺 𝟐 ) = 𝑄 𝟏 (𝑷, 𝜺 𝟏 , 𝜺 𝟐 ) = 𝑄 𝒑 (𝑷, 𝝐 𝟏 𝟏/𝒑 ,𝜺 𝟐 𝟏/𝒑 )

𝐿 0 𝐿 1 Testing Monotonicity Upper bound 𝑂 (log 𝒏/𝝐 ) 𝑂(1/𝝐) Line (𝐷=[𝒏]) 𝐿 0 𝐿 1 Upper bound 𝑂 (log 𝒏/𝝐 ) [Ergun, Kannan, Kumar, Rubinfeld, Viswanathan’00] 𝑂(1/𝝐) Lower bound Ω(log 𝒏/𝝐 ) [Fischer’04] Ω(1/𝝐)

Monotonicity Domain D= [𝑛] 𝑑 (vertices of 𝑑-dim hypercube) (1,1,1) (𝑑,𝑑,𝑑) Domain D= [𝑛] 𝑑 (vertices of 𝑑-dim hypercube) A function 𝑓:𝐷→ℝ is monotone if increasing a coordinate of 𝑥 does not decrease 𝑓 𝑥 . Special case 𝑑=1 𝑓:[𝑛]→ℝ is monotone ⇔ 𝑓 1 ,…𝑓(𝑛) is sorted. One of the most studied properties in property testing [Ergün Kannan Kumar Rubinfeld Viswanathan , Goldreich Goldwasser Lehman Ron, Dodis Goldreich Lehman Raskhodnikova Ron Samorodnitsky, Batu Rubinfeld White, Fischer Lehman Newman Raskhodnikova Rubinfeld Samorodnitsky, Fischer, Halevy Kushilevitz, Bhattacharyya Grigorescu Jung Raskhodnikova Woodruff, ..., Chakrabarty Seshadhri, Blais, Raskhodnikova Yaroslavtsev, Chakrabarty Dixit Jha Seshadhri, …]

Monotonicity: Key Lemma M = class of monotone functions Boolean slicing operator 𝒇 𝒚 :𝐷→{0,1} 𝒇 𝒚 𝑥 =1, if 𝒇 𝑥 ≥ 𝒚, 𝒇 𝒚 𝑥 =0, otherwise. Theorem: 𝑑𝑖𝑠 𝑡 1 𝒇, 𝑀 = ∫ 0 1 𝑑𝑖𝑠 𝑡 0 𝒇 𝒚 , 𝑀 𝑑𝒚

Proof sketch: slice and conquer Closest monotone function with minimal 𝑳 𝟏 -norm is unique (can be denoted as an operator 𝑀 𝒇 1 ). 𝑓 −𝑔 1 = 0 1 𝑓 𝒚 − 𝑔 𝒚 𝑑𝒚 𝑀 𝒇 1 and 𝒇 𝒚 commute: 𝑀 𝒇 1 𝒚 = 𝑀 1 ( 𝒇 𝒚 ) 𝑑𝑖𝑠 𝑡 1 𝑓,𝑀 = 𝑓 − 𝑀 𝑓 1 1 |𝐷| = 0 1 𝑓 𝒚 − ( 𝑀 𝑓 1 ) 𝒚 1 𝑑𝒚 𝐷 = = 0 1 𝑓 𝒚 − 𝑀 (𝑓 𝒚 ) 1 1 𝑑𝒚 𝐷 = 0 1 𝑑𝑖𝑠 𝑡 0 𝑓 𝒚 ,𝑀 𝑑𝒚 2) 3) 1)

𝐿 1 -Testers from Boolean Testers Thm: A nonadaptive, 1-sided error 𝐿 0 -test for monotonicity of 𝑓:𝐷→{0,1} is also an 𝐿 1 -test for monotonicity of 𝑓:𝐷→[0,1]. Proof: A violation (𝑥,𝑦): A nonadaptive, 1-sided error test queries a random set 𝑄⊆𝐷 and rejects iff 𝑄 contains a violation. If 𝑓:𝐷→[0,1] is monotone, 𝑄 will not contain a violation. If 𝑑 1 𝑓,𝑀 ≥𝜀 then ∃ 𝒕 ∗ : 𝑑 0 𝒇 (𝒕 ∗ ) , 𝑀 ≥𝜺 W.p. ≥2/3, set 𝑄 contains a violation (𝑥,𝑦) for 𝒇 (𝒕 ∗ ) 𝒇 ( 𝒕 ∗ ) 𝑥 =1, 𝒇 ( 𝒕 ∗ ) 𝑦 =0 ⇓ 𝒇 𝑥 >𝒇 𝑦 For Boolean functions 𝑂(1/𝜖) sample is enough 𝒇(𝒙) > 𝒇(𝒚)

Our Results: Testing Monotonicity Hypergrid (𝐷= 𝒏 𝒅 ) 2 𝑂 𝒅 /𝜖 adaptive tester for Boolean functions 𝐿 0 𝐿 1 Upper bound 𝑂 𝐝 log 𝒏 𝝐 [Dodis et al. ’99,…, Chakrabarti, Seshadhri ’13] 𝑂 𝒅 𝝐 log 𝒅 𝝐 Lower bound Ω 𝐝 log 𝒏 𝝐 [Dodis et al.’99…, Chakrabarti, Seshadhri ’13] Ω 1 𝝐 log 1 𝝐 Non-adaptive 1-sided error

Testing Monotonicity of 𝒏 𝒅 →{0,1} 𝑒 𝒊 =(0…1…0) = 𝒊-th unit vector. For 𝒊∈ 𝒅 , 𝛼∈ 𝒏 𝒅 where 𝛼 𝒊 =0 an axis-parallel line along dimension 𝒊 : 𝛼 + 𝑥 𝒊 𝑒 𝒊 𝑥 𝒊 ∈[𝒏]} 𝐿 𝒏,𝒅 = set of all 𝒅 𝒏 𝒅−1 axis-parallel lines Dimension reduction for 𝒇: 𝒏 𝒅 → 0,1 [Dodis et al.’99]: 𝐸 ℓ∼ 𝐿 𝒏,𝒅 𝑑𝑖𝑠𝑡 𝒇 ​ ℓ ,𝑀 ≥ 𝑑𝑖𝑠𝑡 𝒇,𝑀 2𝒅 If 𝑑𝑖𝑠𝑡 𝒇 ​ ℓ ,𝑀 ≥𝜹 => 𝑂 1 𝜹 -sample detects a violation

Testing Monotonicity on 𝒏 𝒅 Dimension reduction for 𝒇: 𝒏 𝒅 →{0,1}[Dodis et al.’99]: 𝐸 ℓ∼ 𝐿 𝒏,𝒅 𝑑𝑖𝑠𝑡 𝒇 ​ ℓ ,𝑀 ≥ 𝑑𝑖𝑠𝑡 𝒇,𝑀 2𝒅 If 𝑑𝑖𝑠𝑡 𝒇 ​ ℓ ,𝑀 ≥𝜹 => 𝑂 1 𝜹 -sample can detect a violation “Inverse Markov”: For r.v. 𝑿∈ 0,1 with E 𝐗 =𝝁 and 𝑐<1 Pr 𝑿≤𝑐𝝁 ≤ 1 − 𝝁 1 −𝑐 𝝁 ⇒ Pr 𝑿≤ 𝝁 2 ≤1 − 𝝁 2 −𝝁 ≤1 − 𝝁 2 Pr 𝑑𝑖𝑠𝑡 𝒇 ​ ℓ ,𝑀 ≥ 𝑑𝑖𝑠𝑡 𝒇,𝑀 𝟒𝒅 ≥ 𝑑𝑖𝑠𝑡 𝒇,𝑀 𝟒𝒅 ⇒𝑂 𝒅 𝟐 𝝐 𝟐 -test [Dodis et al.] 𝑂 𝒅 𝝐 log 2 𝒅 𝝐 via “Levin’s economical work investment strategy” (used in other papers for testing connectedness of a graph, properties of images, etc.)

Testing Monotonicity on 𝒏 𝒅 “Discretized Inverse Markov” For r.v. 𝑿∈ 0,1 with E 𝐗 =𝝁≤ 1 2 and 𝒕=3 log 1/𝝁 ∃ 𝒋∈ 𝒕 : Pr 𝑿≥ 2 −𝒋 ≥ 2 𝒋 𝝁 4 For each 𝒊∈[𝒕] pick 𝑂 1 𝝁 2 𝒊 samples of size 𝑂( 2 𝒊 ) => complexity 𝑂 1 𝝁 log 1 𝝁 For the right value j the test rejects with constant probability 𝝁= 𝐸 ℓ∼ 𝐿 𝒏,𝒅 𝑑𝑖𝑠𝑡 𝒇 ​ ℓ ,𝑀 ≥ 𝑑𝑖𝑠𝑡 𝒇,𝑀 2𝒅 => 𝑂 𝒅 𝝐 log 𝒅 𝝐 -test

Distance Approximation and Tolerant Testing Approximating 𝑳 𝟏 -distance to monotonicity ±𝜹 𝒘.𝒑. ≥𝟐/𝟑 𝑓 𝐿 0 𝐿 1 𝒏 →[0,1] polylog 𝒏 ⋅ 𝟏 𝜹 𝑶 𝟏/𝜹 [Saks Seshadhri 10] Θ 𝟏 𝜹 𝟐 Sublinear algorithm for isotonic regression Improves 𝑂 𝟏 𝜹 𝟐 adaptive distance approximation of [Fattal,Ron’10] for Boolean functions Time complexity of tolerant 𝐿 1 -testing for monotonicity is O 𝜺 𝟐 ( 𝜺 𝟐 − 𝜺 𝟏 ) 𝟐 Better dependence than what follows from distance appoximation for 𝝐 𝟐 ≪1

Distance Approximation 𝑓: 𝑛 → 0,1 Theorem: with constant probability over the choice of a random sample S of size O 1 𝛿 2 : 𝑑𝑖𝑠 𝑡 1 𝑓 ​ 𝑺 ,𝑀 −𝑑𝑖𝑠 𝑡 1 𝑓,𝑀 <𝛿 Implies an O 1 𝜖 2 − 𝜖 1 2 tolerant tester by setting 𝛿= 𝜖 2 − 𝜖 1 3 𝑑𝑖𝑠 𝑡 1 𝒇, 𝑀 = ∫ 0 1 𝑑𝑖𝑠 𝑡 0 𝒇 𝒚 , 𝑀 𝑑𝒚 Suffices: ∀𝒚: 𝑑𝑖𝑠 𝑡 0 𝒇 𝒚 ​ 𝑺 ,𝑀 −𝑑𝑖𝑠 𝑡 0 𝒇 𝒚 ,𝑀 <𝛿 Improves previous 𝑂 (1/ 𝛿 2 ) algorithm [Fattal, Ron’10]

Distance Approximation For 𝒇:[𝑛]→ 0,1 violation graph 𝑮 𝒇 𝑛 , 𝐸 : edge ( 𝑥 1 , 𝑥 2 ) if 𝑥 1 ≤ 𝑥 2 , 𝑓 𝑥 1 =1, 𝑓 𝑥 2 =0 MM(G) = maximum matching VC(G) = minimum vertex cover 𝑑𝑖𝑠 𝑡 0 𝒇,𝑀 = 𝑴𝑴 𝐺 𝑓 |𝐷| = 𝑽𝑪 𝐺 𝑓 |𝐷| [Fischer et al.’02] 𝑑𝑖𝑠 𝑡 0 𝒇 ​ 𝑆 ,𝑀 = 𝑴𝑴 𝐺 𝑓 ​ 𝑆 𝑆 = 𝑽𝑪 𝐺 𝑓 ​ 𝑆 |𝑆|

𝑑𝑖𝑠 𝑡 0 𝒇 ​ 𝑺 ,𝑀 −𝑑𝑖𝑠 𝑡 0 𝒇,𝑀 <O 1 𝑺 Define: 𝒀 𝑺 = 𝑽𝑪 𝒇 ∩𝑺 𝑺 𝑑𝑖𝑠 𝑡 0 𝒇 ​ 𝑺 ,𝑀 = 𝑽𝑪 𝑓 ​ 𝑺 |𝑺| ≤ 𝑽𝑪 𝑓 ∩𝑺 𝑺 = 𝒀(𝑺) 𝒀(𝑺) has hypergeometric distribution: 𝐸 𝒀 𝑺 = 𝑽𝑪 𝒇 𝐷 =𝑑𝑖𝑠 𝑡 0 𝒇,𝑀 𝑉𝑎𝑟 𝑌 𝑺 ≤ 𝑺 𝑉 𝐶 𝑓 𝐷 𝑺 2 = 𝑑𝑖𝑠 𝑡 0 𝑓,𝑀 𝑺 ≤ 1 |𝑺|

Experiments Data: Apple stock price data (2005-2015) from Google Finance Left: 𝐿 1 -isotonic regression Right: multiplicative error vs. sample size

𝐿 1 -Testers for Other Properties Via combinatorial characterization of 𝐿 1 -distance to the property Lipschitz property 𝒇: 𝒏 𝒅 →[0,1]: Θ 𝒅 𝜖 Via (implicit) proper learning: approximate in 𝐿 1 up to error 𝝐, test approximation on a random 𝑂(1/𝜖)-sample Convexity 𝒇: 𝒏 𝒅 →[0,1]: O 𝝐 − 𝒅 2 + 1 𝝐 (tight for 𝒅≤2) Submodularity 𝒇: 0,1 𝒅 → 0,1 2 𝑂 1 𝝐 +𝑝𝑜𝑙𝑦 1 𝝐 log 𝒅 [Feldman, Vondrak 13, …]

𝐿 𝑝 -Testing for Convex Optimization Theory: Convergence rates of gradient descent methods depends on: Convexity / strong convexity constant Lipschitz constant of the derivative Practice: Q: How to pick learning rate in ML packages? A: Set 0.01 and hope it converges fast Even non-tolerant 𝐿 𝑝 -testers can be used to sanity check convexity/Lipschitzness

A lot of open problems! 𝐿 𝑝 -Testing Fourier sparsity [Backurs, Blais, Kapralov, Onak, Y.] Eric Price: Hey, I can do this better!

Tolerant testers for higher dimensions? Open Problems Our complexity for 𝐿 𝑝 -testing convexity grows exponentially with d Is there an 𝐿 𝑝 -testing algorithm for convexity with subexponential dependence on the dimension? Only have tolerant monotonicity for 𝑑=1,2. Tolerant testers for higher dimensions?