Presentation is loading. Please wait.

Presentation is loading. Please wait.

Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi

Similar presentations


Presentation on theme: "Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi"β€” Presentation transcript:

1 Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi
LP, SOCP, and Optimization-Free Approaches to Polynomial Optimization and Difference of Convex Programming Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi

2 Two problems of interest
Lower bounds for polynomial optimization problems Writing a polynomial as a difference of two convex polynomials Can we provide lower bounds on the polynomial optimization problem (POP) min π‘₯∈ ℝ 𝑛 𝑝(π‘₯) s.t. 𝑔 𝑖 π‘₯ β‰₯0, 𝑖=1,…,π‘š where 𝑝, 𝑔 𝑖 , 𝑖=1,…,π‘š are polynomials? Given a polynomial 𝑓 π‘₯ , can we find convex polynomials 𝑔 π‘₯ and β„Ž(π‘₯) such that 𝑓 π‘₯ =𝑔 π‘₯ βˆ’β„Ž π‘₯ ? First part of the talk Second part of the talk

3 Part I: Computing lower bounds for POPs

4 Lower bounds on POPs: applications (1/2)
Polynomial optimization problems min π‘₯∈ ℝ 𝑛 𝑝(π‘₯) 𝑠.𝑑. 𝑔 𝑖 π‘₯ β‰₯0 Optimal power flow problem Economics and game theory Sensor network localization

5 Lower bounds on POPs: applications (2/2)
Infeasibility of a set of polynomial inequalities The system 𝑔 𝑖 π‘₯ β‰₯0, 𝑖=1,…,π‘š is infeasible. 0 is a strict lower bound on min π‘₯ βˆ’π‘” 1 (π‘₯) s.t. 𝑔 2 π‘₯ β‰₯0,…, 𝑔 π‘š π‘₯ β‰₯0 ⇔ If 0 is not a strict upper bound on g_1(x) then it means that there exists x such that g2(x)>=0,…,gm(x)>=0 and g1(x)>=0 which means the system is feasible Automated theorem proving Optimality certificates for discrete optimization

6 Lower bounds on POPs: the SOS approach (1/2)
A polynomial 𝑝 is a sum of squares (SOS) if it can be written as 𝑝 π‘₯ = 𝑖 π‘ž 𝑖 2 (π‘₯) , where π‘ž 𝑖 are polynomials. Example: Sufficient condition for nonnegativity. A polynomial 𝑝(π‘₯) of degree 2d is SOS if and only if βˆƒπ‘„β‰½0 such that where 𝑧= 1, π‘₯ 1 ,…, π‘₯ 𝑛 , π‘₯ 1 π‘₯ 2 ,…, π‘₯ 𝑛 𝑑 𝑇 is the vector of monomials of degree up to 𝑑.

7 Lower bounds on POPs: the SOS approach (2/2)
max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύβ‰₯0, βˆ€π‘₯∈ π‘₯ 𝑔 𝑖 π‘₯ β‰₯0, 𝑖=1,…,π‘š} min π‘₯ 𝑝(π‘₯) 𝑠.𝑑. 𝑔 𝑖 π‘₯ β‰₯0, 𝑖=1,…,π‘š = opt. val. 𝜸 βˆ— Under compactness assumptions, for large enough degree = opt. val. max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ + 𝑖 𝜎 𝑖 π‘₯ β‹… 𝑔 𝑖 π‘₯ , where 𝜎 𝑖 , 𝑖=0,…,π‘š, are SOS. For fixed degrees of the SOS polynomials, solving this problem is an SDP This is an SOS certificate that 𝛾 is a lower bound for POP. There are many other such SOS certificates.

8 The first part of the talk
Goal: avoid using SOS polynomials and SDP to compute lower bounds for POPs LP and SOCP-based alternatives to SOS polynomials (applies to any SOS-based certificate) Using dsos/sdsos polynomials and improvements New certificate that 𝛾 is a lower bound on the POP that only requires checking if coefficients of a given polynomial are β‰₯0 Providing a new certificate that does not use optimization at all max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ + 𝑖 𝜎 𝑖 π‘₯ β‹… 𝑔 𝑖 π‘₯ , where 𝜎 𝑖 , 𝑖=0,…,π‘š, are SOS. max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ + 𝑖 𝜎 𝑖 π‘₯ β‹… 𝑔 𝑖 π‘₯ , where 𝜎 𝑖 , 𝑖=0,…,π‘š, are SOS. dsos/sdsos + improvements New certificate

9 Part I: Computing lower bounds for POPs
Using dsos/sdsos polynomials and improvements Providing an optimization-free certificate for lower bounds on POPs

10 LP and SOCP alternatives to sos: dsos and sdsos
Sum of squares (sos) 𝑝 π‘₯ =𝑧 π‘₯ 𝑇 𝑄𝑧 π‘₯ , 𝑄≽0 SDP DD cone ≔ 𝑸 𝑸 π’Šπ’Š β‰₯ 𝒋 𝑸 π’Šπ’‹ , βˆ€π’Š} PSD cone≔ 𝑸 π‘Έβ‰½πŸŽ} SDD cone ≔ 𝑸 βˆƒ diagonal 𝑫 with 𝑫 π’Šπ’Š >𝟎 s.t. 𝑫𝑸𝑫 𝒅𝒅} Diagonally dominant sum of squares (dsos) 𝑝 π‘₯ =𝑧 π‘₯ 𝑇 𝑄𝑧 π‘₯ , 𝑄 π‘‘π‘–π‘Žπ‘”π‘œπ‘›π‘Žπ‘™π‘™π‘¦ π‘‘π‘œπ‘šπ‘–π‘›π‘Žπ‘›π‘‘ (dd) LP Scaled diagonally dominant sum of squares (sdsos) 𝑝 π‘₯ =𝑧 π‘₯ 𝑇 𝑄𝑧 π‘₯ , 𝑄 π‘ π‘π‘Žπ‘™π‘’π‘‘ π‘‘π‘–π‘Žπ‘”π‘œπ‘›π‘Žπ‘™π‘™π‘¦ π‘‘π‘œπ‘šπ‘–π‘›π‘Žπ‘›π‘‘ (sdd) SOCP Ahmadi, Majumdar

11 where 𝜎 𝑖 , 𝑖=0,…,π‘š, are dsos/sdsos.
LP and SOCP alternatives to sos: dsos and sdsos Problem: computing lower bounds for POPs max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ + 𝑖 𝜎 𝑖 π‘₯ β‹… 𝑔 𝑖 π‘₯ , where 𝜎 𝑖 , 𝑖=0,…,π‘š, are sos. max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ + 𝑖 𝜎 𝑖 π‘₯ β‹… 𝑔 𝑖 π‘₯ , where 𝜎 𝑖 , 𝑖=0,…,π‘š, are dsos/sdsos. β‰₯ opt. val. b Example: For a parametric family of polynomials: 𝑝 π‘₯ 1 , π‘₯ 2 =2 π‘₯ π‘₯ 2 4 +π‘Ž π‘₯ 1 3 π‘₯ 2 +(1βˆ’π‘Ž) π‘₯ 1 2 π‘₯ 2 2 +𝑏 π‘₯ 1 π‘₯ 2 3 a

12 Improvements on dsos and sdsos (1/2)
Replacing sos polynomials by dsos/sdsos polynomials: +: fast bounds - : not always as good quality (compared to sos) Iteratively construct a sequence of improving LP/SOCPs Initialization: Start with the dsos/sdsos polynomials Column Generation method

13 Improvements on dsos/sdsos (2/2)
Problem of interest: computing lower bounds for POPs Example considered : computing lower bounds for unconstrained POPs Compactness assumptions + large degree max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ + 𝑖 𝜎 𝑖 π‘₯ β‹… 𝑔 𝑖 π‘₯ , where 𝜎 𝑖 , 𝑖=0,…,π‘š, are SOS. min π‘₯ 𝑝(π‘₯) 𝑠.𝑑. 𝑔 𝑖 π‘₯ β‰₯0 = opt. val. max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= 𝜎 0 π‘₯ , where 𝜎 0 is SOS. β‰₯ opt. val. min π‘₯ 𝑝(π‘₯)

14 Method 2: Column generation (1/3)
Focus on LP-based version of this method (SOCP is similar). Sos problem max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ SOS Replace with Dsos problem max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ dsos Dsos problem max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ=𝒛 𝒙 𝑻 𝑸𝒛 𝒙 , 𝑸 𝒅𝒅 Two different ways of characterizing 𝑸 𝒅𝒅: β‘‘ 𝑄 dd β‡”βˆƒ 𝛼 𝑖 β‰₯0 s.t. 𝑄= 𝑖 𝛼 𝑖 𝑣 𝑖 𝑣 𝑖 𝑇 , where 𝑣 𝑖 fixed vector with at most two nonzero components =Β±1 β‘  𝑄 dd ⇔ 𝑄 𝑖𝑖 β‰₯ 𝑗 |𝑄 𝑖𝑗 | , βˆ€ 𝑖 [In collaboration with Sanjeeb Dash, IBM Research]

15 Method 2: Column generation (2/3)
Dsos problem max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ= π’Š 𝜢 π’Š 𝒗 π’Š 𝑻 𝒛 𝒙 𝟐 , 𝜢 π’Š β‰₯𝟎 Dsos problem max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ=𝒛 𝒙 𝑻 𝑸𝒛 𝒙 , 𝑸 𝒅𝒅 Using β‘‘ Idea behind the algorithm: Expand feasible space at each iteration by adding a new vector 𝑣 and variable 𝛼 max 𝛾 𝛾 𝒑 𝒙 βˆ’πœΈ= π’Š 𝜢 π’Š 𝒗 π’Š 𝑻 𝒛 𝒙 𝟐 +𝜢 𝒗 𝑻 𝒛 𝒙 𝟐 𝜢 π’Š β‰₯𝟎, 𝒂β‰₯𝟎 Question: How to pick 𝒗? Use the dual!

16 Method 2: Column Generation (3/3)
DSOS + 5 iterations SDSOS + 5 iterations Example: minimizing a degree-4 polynomial 𝒏=πŸπŸ“ 𝒏=𝟐𝟎 𝒏=πŸπŸ“ 𝒏=πŸ‘πŸŽ 𝒏=πŸ’πŸŽ bd t(s) DSOS -10.96 0.38 -18.01 0.74 -26.45 15.51 -36.52 7.88 -62.30 10.68 𝐷𝑆𝑂 𝑆 π‘˜ -5.57 31.19 -9.02 471.39 -20.08 600 -32.28 -35.14 SOS -3.26 5.60 -3.58 82.22 -3.71 NA

17 Part I: Avoiding SDP to compute lower bounds on POPs
Using dsos/sdsos polynomials and improvements Providing an optimization-free certificate for lower bounds on POPs

18 An optimization free method for lower bounds (1/3)
max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύβ‰₯0, βˆ€π‘₯∈ π‘₯ 𝑔 𝑖 π‘₯ β‰₯0, 𝑖=1,…,π‘š} min π‘₯∈ ℝ 𝑛 𝑝(π‘₯) 𝑠.𝑑. 𝑔 𝑖 π‘₯ β‰₯0, 𝑖=1,…,π‘š = opt. val. 2𝑑= maximum degree of 𝑝, 𝑔 𝑖 Under compactness assumptions, i.e., π‘₯ 𝑔 𝑖 π‘₯ β‰₯0}βŠ†π΅(0,𝑅) = opt. val. For large enough π‘Ÿ sup 𝛾 𝛾 s.t. 𝑓 𝛾 𝑣 2 βˆ’ 𝑀 2 βˆ’ 1 π‘Ÿ 𝑖 (𝑣 𝑖 2 βˆ’ 𝑀 𝑖 2 ) 2 𝑑 + 1 2π‘Ÿ 𝑖 𝑣 𝑖 4 + 𝑀 𝑖 𝑑 β‹… 𝑖 𝑣 𝑖 2 + 𝑖 𝑀 𝑖 π‘Ÿ 2 has nonnegative coefficients, where 𝑓 𝛾 is a form in 𝑛+π‘š+3 variables and of degree 4𝑑 which explicitely depends on 𝑝, 𝑔 𝑖 and 𝑅 Theorem:

19 An optimization free method for lower bounds (2/3)
𝛾 is a strict lower bound on the POP ⇔ βˆƒπ‘Ÿβˆˆβ„• s.t. 𝒇 𝜸 𝒗 𝟐 βˆ’ π’˜ 𝟐 βˆ’ 𝟏 𝒓 π’Š (𝒗 π’Š 𝟐 βˆ’ π’˜ π’Š 𝟐 ) 𝟐 𝒅 + 𝟏 πŸπ’“ π’Š 𝒗 π’Š πŸ’ + π’˜ π’Š πŸ’ 𝒅 β‹… 𝑖 𝑣 𝑖 2 + 𝑖 𝑀 𝑖 π‘Ÿ 2 has β‰₯0 coefficients Proof sketch of the theorem: 𝛾 is a strict lower bound on the POP ⇔ βˆƒπ‘Ÿβˆˆβ„• s.t. 𝒇 𝜸 𝒗 𝟐 βˆ’ π’˜ 𝟐 βˆ’ 1 π‘Ÿ 𝑖 (𝑣 𝑖 2 βˆ’ 𝑀 𝑖 2 ) 2 𝑑 + 𝟏 πŸπ’“ π’Š 𝒗 π’Š πŸ’ + π’˜ π’Š πŸ’ 𝒅 β‹… 𝑖 𝑣 𝑖 2 + 𝑖 𝑀 𝑖 π‘Ÿ 2 has β‰₯0 coefficients 𝛾 is a strict lower bound on the POP ⇔ βˆƒπ‘Ÿβˆˆβ„• s.t. 𝒇 𝜸 𝒗 𝟐 βˆ’ π’˜ 𝟐 βˆ’ 1 π‘Ÿ 𝑖 (𝑣 𝑖 2 βˆ’ 𝑀 𝑖 2 ) 2 𝑑 + 1 2π‘Ÿ 𝑖 𝑣 𝑖 4 + 𝑀 𝑖 𝑑 β‹… 𝑖 𝑣 𝑖 2 + 𝑖 𝑀 𝑖 π‘Ÿ 2 has β‰₯0 coefficients 𝛾 is a strict lower bound on the POP ⇔ βˆƒπ‘Ÿβˆˆβ„• s.t. 𝑓 𝛾 𝑣 2 βˆ’ 𝑀 2 βˆ’ 1 π‘Ÿ 𝑖 (𝑣 𝑖 2 βˆ’ 𝑀 𝑖 2 ) 2 𝑑 + 1 2π‘Ÿ 𝑖 𝑣 𝑖 4 + 𝑀 𝑖 𝑑 β‹… 𝑖 𝑣 𝑖 2 + 𝑖 𝑀 𝑖 π‘Ÿ 2 has β‰₯0 coefficients 𝛾 is a lower bound on the POP β‡’ 𝑓 𝛾 is pd Result by Polya: 𝑓 even and pd β‡’βˆƒ π‘Ÿβˆˆβ„• such that 𝑓 𝑧 β‹… 𝑖 𝑧 𝑖 2 π‘Ÿ has nonnegative coefficients. Make 𝑓 𝛾 (𝑧) even by considering 𝒇 𝜸 𝒗 𝟐 βˆ’ π’˜ 𝟐 . We lose positive definiteness of 𝑓 𝛾 with this transformation. Add the positive definite term 𝟏 πŸπ’“ π’Š 𝒗 π’Š πŸ’ + π’˜ π’Š πŸ’ ) to 𝑓 𝛾 ( 𝑣 2 βˆ’ 𝑀 2 ) to make it positive definite. Apply Polya’s result. The term βˆ’ 𝟏 𝒓 π’Š (𝒗 π’Š 𝟐 βˆ’ π’˜ π’Š 𝟐 ) 𝟐 𝒅 ensures that the converse holds as well.

20 An optimization free method for lower bounds (3/3)
How to use this method in practice? Given a polynomial optimization problem: min 𝑝(π‘₯) , s.t. 𝑔 𝑖 π‘₯ β‰₯0, 𝑅 such that π‘₯ 𝑔 𝑖 π‘₯ β‰₯0}βŠ†π΅(0,𝑅), an accuracy 𝝐 for the optimal value of the POP 𝑝 βˆ— Expand and check nonnegativity of the coefficients of the polynomial 𝑓 𝛾 𝑣 2 βˆ’ 𝑀 2 βˆ’ 1 π‘Ÿ 𝑖 (𝑣 𝑖 2 βˆ’ 𝑀 𝑖 2 ) 2 𝑑 + 1 2π‘Ÿ 𝑖 𝑣 𝑖 4 + 𝑀 𝑖 𝑑 β‹… 𝑖 𝑣 𝑖 2 + 𝑖 𝑀 𝑖 π‘Ÿ 2 Have we reached accuracy πœ– in bisection? Start bisection on 𝛾 by taking 𝛾= π‘ˆ 0 + 𝐿 0 2 π‘Ÿ=1 NB: 𝐿 0 ≀ 𝑝 βˆ— ≀ π‘ˆ 0 No: keep doing bisection Yes … π‘Ÿ=2

21 Part II: Writing a polynomial as a difference of two convex polynomials

22 Difference of convex decomposition: Applications
Can be used for difference of convex programs, i.e., problems of the form: min 𝑓 0 (π‘₯) 𝑠.𝑑. 𝑓 𝑖 π‘₯ ≀0 where 𝑓 𝑖 π‘₯ ≔ 𝑔 𝑖 π‘₯ βˆ’ β„Ž 𝑖 π‘₯ , 𝑔 𝑖 , β„Ž 𝑖 convex. Applications: Machine Learning (Sparse PCA, Kernel selection, feature selection in SVM) Studied for quadratics, polynomials nice to study question computationally Hiriart-Urruty, 1985 Tuy, 1995

23 Difference of Convex (dc) decomposition
Difference of convex (dc) decomposition: given a polynomial 𝑓, find 𝑔 and β„Ž such that 𝒇=π’ˆβˆ’π’‰, where 𝑔,β„Ž convex polynomials. Questions: Does such a decomposition always exist? Can I obtain such a decomposition efficiently? Is this decomposition unique?

24 Existence of dc decomposition (1/5)
Recall: Theorem: Any polynomial can be written as the difference of two dsos-convex polynomials. Corollary: Any polynomial can be written as the difference of two sdsos-convex/sos-convex/convex polynomials. 𝑦 𝑇 𝐻 𝑓 π‘₯ 𝑦 sos SDP SOS-convexity 𝑦 𝑇 𝐻 𝑝 π‘₯ 𝑦 sdsos SOCP SDSOS-convexity 𝑦 𝑇 𝐻 𝑝 π‘₯ 𝑦 dsos LP DSOS-convexity 𝑓(π‘₯) convex 𝑦 𝑇 𝐻 𝑓 π‘₯ 𝑦β‰₯0, βˆ€π‘₯,π‘¦βˆˆ ℝ 𝑛 ⇐ ⇔ ⇐

25 Existence of dc decomposition (2/5)
Lemma: Let 𝐾 be a full dimensional cone in a vector space 𝐸. Then any π‘£βˆˆ 𝐸 can be written as 𝑣= π‘˜ 1 βˆ’ π‘˜ 2 , π‘˜ 1 , π‘˜ 2 ∈𝐾. Proof sketch: =:π‘˜β€² βˆƒ 𝛼<1 such that 1βˆ’π›Ό 𝑣+π›Όπ‘˜βˆˆπΎ E K ⇔𝑣= 1 1βˆ’π›Ό π‘˜ β€² βˆ’ 𝛼 1βˆ’π›Ό π‘˜ To change π’Œ π’Œβ€² 𝒗 π‘˜ 1 ∈𝐾 π‘˜ 2 ∈𝐾

26 Existence of dc decomposition (3/5)
Here, 𝐸={polynomials of degree d, in n variables}, 𝐾={dsos-convex polynomials of degree d and in n variables }. Remains to show that 𝐾 is full dimensional: the polynomial in the interior is constructed inductively on 𝑛 with appropriately scaled coefficients. This shows that a decomposition can be obtained efficiently: 𝒇=π’ˆβˆ’π’‰, π’ˆ,𝒉 s/d/sos-convex solving is an LP/SOCP/SDP.

27 Uniqueness of dc decomposition
Dc decomposition: given a polynomial 𝑓, find convex polynomials 𝑔 and β„Ž such that 𝒇=π’ˆβˆ’π’‰. Questions: Does such a decomposition always exist? Can I obtain such a decomposition efficiently? Is this decomposition unique? οƒΌ Yes οƒΌ Through s/d/sos-convexity Alternative decompositions 𝑓 π‘₯ = 𝑔 π‘₯ +𝑝 π‘₯ βˆ’ β„Ž π‘₯ +𝑝 π‘₯ 𝑝(π‘₯) convex Initial decomposition x𝑓 π‘₯ =𝑔 π‘₯ βˆ’β„Ž(π‘₯) β€œBest decomposition?”

28 Convex-Concave Procedure (CCP)
Heuristic for minimizing DC programming problems. Idea: Input π‘˜β‰”0 x π‘₯ 0 , initial point 𝑓 𝑖 = 𝑔 𝑖 βˆ’ β„Ž 𝑖 , 𝑖=0,…,π‘š Convexify by linearizing 𝒉 x 𝒇 π’Š π’Œ 𝒙 = 𝑔 𝑖 π‘₯ βˆ’( β„Ž 𝑖 π‘₯ π‘˜ +𝛻 β„Ž 𝑖 π‘₯ π‘˜ 𝑇 π‘₯βˆ’ π‘₯ π‘˜ ) Solve convex subproblem Take π‘₯ π‘˜+1 to be the solution of min 𝑓 0 π‘˜ π‘₯ 𝑠.𝑑. 𝑓 𝑖 π‘˜ π‘₯ ≀0, 𝑖=1,…,π‘š convex convex affine π‘˜β‰”π‘˜+1 𝒇 π’Š π’Œ 𝒙 𝒇 π’Š (𝒙)

29 Convex-Concave Procedure (CCP)
Toy example: min π‘₯ 𝑓 π‘₯ , where 𝑓 π‘₯ ≔𝑔 π‘₯ βˆ’β„Ž(π‘₯) Convexify 𝑓 π‘₯ to obtain 𝑓 0 (π‘₯) Initial point: π‘₯ 0 =2 Minimize 𝑓 0 (π‘₯) and obtain π‘₯ 1 Reiterate π‘₯ ∞ π‘₯ 3 π‘₯ 4 π‘₯ 2 π‘₯ 1 π‘₯ 0 π‘₯ 0

30 Picking the β€œbest” decomposition for CCP
Algorithm Linearize 𝒉 𝒙 around a point π‘₯ π‘˜ to obtain convexified version of 𝒇(𝒙) Algorithm Linearize 𝒉 𝒙 around a point π‘₯ π‘˜ to obtain convexified version of 𝒇(𝒙) Algorithm Linearize 𝒉 𝒙 around a point π‘₯ π‘˜ to obtain convexified version of 𝒇(𝒙) Idea Pick β„Ž π‘₯ such that it is as close as possible to affine around π‘₯ π‘˜ Idea Pick β„Ž π‘₯ such that it is as close as possible to affine around π‘₯ π‘˜ Mathematical translation Minimize curvature of β„Ž

31 Undominated decompositions (1/2)
Definition: g ,β„Žβ‰”π‘”βˆ’f is an undominated decomposition of 𝑓 if no other decomposition of 𝑓 can be obtained by subtracting a (nonaffine) convex function from 𝑔. π’ˆ 𝒙 = 𝒙 πŸ’ + 𝒙 𝟐 , 𝒉 𝒙 =πŸ’ 𝒙 𝟐 +πŸπ’™βˆ’πŸ Convexify around π‘₯ 0 =2 to get 𝒇 𝟎 𝒙 𝒇 𝒙 = 𝒙 πŸ’ βˆ’πŸ‘ 𝒙 𝟐 +πŸπ’™βˆ’πŸ Cannot substract something convex from g and get something convex again. DOMINATED BY π’ˆ β€² 𝒙 = 𝒙 πŸ’ , 𝒉 β€² 𝒙 =πŸ‘ 𝒙 𝟐 +πŸπ’™βˆ’πŸ Convexify around π‘₯ 0 =2 to get 𝒇 πŸŽβ€² 𝒙 If π’ˆβ€² dominates π’ˆ then the next iterate in CCP obtained using π’ˆ β€² always beats the one obtained using π’ˆ.

32 Undominated decompositions (2/2)
Theorem: Given a polynomial 𝑓, consider min 1 𝐴 𝑛 𝑆 π‘›βˆ’1 π‘‡π‘Ÿ 𝐻 𝑔 π‘‘πœŽ , (where 𝐴 𝑛 = 2 πœ‹ 𝑛/2 Ξ“(𝑛/2) ) s.t. 𝑓=π‘”βˆ’β„Ž, 𝑔 convex, β„Ž convex Any optimal solution is an undominated dcd of 𝑓 (and an optimal solution always exists). Theorem: If 𝑓 has degree 4, it is strongly NP-hard to solve (⋆). Idea: Replace 𝑓=π‘”βˆ’β„Ž, 𝑔, β„Ž convex by 𝑓=π‘”βˆ’β„Ž, 𝑔,β„Ž sos-convex. (⋆) 𝑔,β„Ž

33 Comparing different decompositions (1/2)
Solving the problem: min 𝐡= π‘₯ π‘₯ ≀𝑅} 𝑓 0 , where 𝑓 0 has 𝑛=8 and 𝑑=4. Decompose 𝑓 0 , run CCP for 4 minutes and compare objective value. Feasibility Undominated min 𝑔,β„Ž 1 𝐴 𝑛 𝑆 π‘›βˆ’1 π‘‡π‘Ÿ 𝐻 𝑔 π‘‘πœŽ 𝑠.𝑑. 𝑓 0 =π‘”βˆ’β„Ž 𝑔,β„Ž sos-convex min g,h 0 s.t. 𝑓 0 =π‘”βˆ’β„Ž 𝑔,β„Ž sos-convex

34 Comparing different decompositions (2/2)
Average over 30 instances Solver: Mosek Computer: 8Gb RAM, 2.40GHz processor Feasibility Undominated Conclusion: Performance of CCP strongly affected by initial decomposition.

35 Main messages (1/2) Computing lower bounds on POPs has many applications. Traditionally, one can obtain lower bounds on POPs using sos polynomials. Here, we presented two different methods to not use sos polynomials. Replace sos polynomials by dsos/sdsos polynomials Improve on these via sequences of LPs and SOCPs Use a new optimization-free certificate to certify lower bounds on POPs (only uses multiplication and coefficient checking)

36 Main messages (2/2) Decomposing a polynomial into the difference of two convex polynomials has long been a problem of interest. A decomposition can be obtained efficiently using sos-convexity and SDP We can also obtain decompositions using LP and SOCP. We show that the decomposition we pick impacts performance of the CCP algorithm, a heuristic to solve DC programs.

37 Thank you for listening
Questions? Want to learn more?

38 Method 1: Cholesky change of basis (1/3)
dd in the β€œright basis” psd but not dd Goal: iteratively improve on basis

39 Method 1: Cholesky change of basis (2/3)
Sos problem max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ SOS Initialize max 𝛾 𝛾 s.t. 𝑝 π‘₯ βˆ’π›Ύ=𝒛 𝒙 𝑻 𝑸𝒛 𝒙 , 𝑸 𝒅𝒅/𝒔𝒅𝒅 Step 1 Replace: π‘ˆ 1 =π‘β„Žπ‘œπ‘™( 𝑄 βˆ— ) Step 1 Replace: π‘ˆ π‘˜ =π‘β„Žπ‘œπ‘™( π‘ˆ π‘˜βˆ’1 𝑇 𝑄 βˆ— π‘ˆ π‘˜βˆ’1 ) Step 2 max 𝛾 𝛾 s.t. 𝑝 π‘₯ βˆ’π›Ύ=𝒛 𝒙 𝑻 𝑼 𝟏 𝑻 𝑸 𝑼 𝟏 𝒛 𝒙 , 𝑸 𝒅𝒅/𝒔𝒅𝒅 Step 2 max 𝛾 𝛾 𝑠.𝑑. 𝑝 π‘₯ βˆ’π›Ύ=𝒛 𝒙 𝑻 𝑼 π’Œ 𝑻 𝑸 𝑼 π’Œ 𝒛 𝒙 , 𝑸 𝒅𝒅/𝒔𝒅𝒅 New basis π’Œβ‰”π’Œ+𝟏 One iteration of this method on a parametric family of polynomials: 𝑝 π‘₯ 1 , π‘₯ 2 =2 π‘₯ π‘₯ 2 4 +π‘Ž π‘₯ 1 3 π‘₯ 2 +(1βˆ’π‘Ž) π‘₯ 1 2 π‘₯ 2 2 +𝑏 π‘₯ 1 π‘₯ 2 3

40 Method 1: Cholesky change of basis (3/3)
Example: min π‘₯ 𝑝(π‘₯) , where 𝑝 is in 4 variables and has degree 4 Lower bound on optimal value

41 Method 2: Column generation (3/4)
PRIMAL DUAL A general SDP max π‘¦βˆˆ ℝ π‘š 𝑏 𝑇 𝑦 𝑠.𝑑. πΆβˆ’ 𝑖=1 π‘š 𝑦 𝑖 𝐴 𝑖 ≽0 𝑺𝑫𝑷 Dual of SDP min π‘‹βˆˆ 𝑆 𝑛 π‘‘π‘Ÿ(𝐢𝑋) 𝑠.𝑑. π‘‘π‘Ÿ 𝐴 𝑖 𝑋 = 𝑏 𝑖 𝑋≽0 𝑳𝑷 𝑺𝑫 𝑷 βˆ— 𝑳 𝑷 βˆ— LP obtained with inner approximation of PSD by DD Dual of LP min π‘‹βˆˆ 𝑆 𝑛 π‘‘π‘Ÿ(𝐢𝑋) 𝑠.𝑑. π‘‘π‘Ÿ 𝐴 𝑖 𝑋 = 𝑏 𝑖 𝑣 𝑖 𝑇 𝑋 𝑣 𝑖 β‰₯0 max π‘¦βˆˆ ℝ π‘š 𝑏 𝑇 𝑦 𝑠.𝑑. πΆβˆ’ 𝑖=1 π‘š 𝑦 𝑖 𝐴 𝑖 = 𝛼 𝑖 𝑣 𝑖 𝑣 𝑖 𝑇 𝛼 𝑖 β‰₯0 Pick 𝒗 s.t. 𝒗 𝑻 𝑿𝒗<𝟎.


Download ppt "Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi"

Similar presentations


Ads by Google