Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal January 31, 2007 Inexact Methods for PDE-Constrained Optimization.

Slides:



Advertisements
Similar presentations
Zhen Lu CPACT University of Newcastle MDC Technology Reduced Hessian Sequential Quadratic Programming(SQP)
Advertisements

Optimization with Constraints
Optimization.
Engineering Optimization
دانشگاه صنعتي اميركبير دانشكده مهندسي پزشكي Constraints in MPC کنترل پيش بين-دکتر توحيدخواه.
Lena Gorelick Joint work with Frank Schmidt and Yuri Boykov Rochester Institute of Technology, Center of Imaging Science January 2013 TexPoint fonts used.
Control Structure Selection for a Methanol Plant using Hysys/Unisim
Optimization Methods TexPoint fonts used in EMF.
Page 1 Page 1 ENGINEERING OPTIMIZATION Methods and Applications A. Ravindran, K. M. Ragsdell, G. V. Reklaitis Book Review.
1 TTK4135 Optimization and control B.Foss Spring semester 2005 TTK4135 Optimization and control Spring semester 2005 Scope - this you shall learn Optimization.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
by Rianto Adhy Sasongko Supervisor: Dr.J.C.Allwright
Separating Hyperplanes
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Numerical Optimization
1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me .
Nonlinear Optimization for Optimal Control
Methods For Nonlinear Least-Square Problems
Engineering Optimization
Efficient Methodologies for Reliability Based Design Optimization
Optimization Methods One-Dimensional Unconstrained Optimization
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
Unconstrained Optimization Problem
Engineering Optimization
Advanced Topics in Optimization
1 Multiple Kernel Learning Naouel Baili MRL Seminar, Fall 2009.
Ch. 9: Direction Generation Method Based on Linearization Generalized Reduced Gradient Method Mohammad Farhan Habib NetLab, CS, UC Davis July 30, 2010.
Tier I: Mathematical Methods of Optimization
Optimization of Linear Problems: Linear Programming (LP) © 2011 Daniel Kirschen and University of Washington 1.
UNCONSTRAINED MULTIVARIABLE
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 9. Optimization problems.
KKT Practice and Second Order Conditions from Nash and Sofer
Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal February 12, 2007 Inexact Methods for PDE-Constrained Optimization.
ENCI 303 Lecture PS-19 Optimization 2
General Nonlinear Programming (NLP) Software
Nonlinear Programming.  A nonlinear program (NLP) is similar to a linear program in that it is composed of an objective function, general constraints,
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
L8 Optimal Design concepts pt D
1 Algorithms and Software for Large-Scale Nonlinear Optimization OTC day, 6 Nov 2003 Richard Waltz, Northwestern University Project I: Large-scale Active-Set.
A comparison between PROC NLP and PROC OPTMODEL Optimization Algorithm Chin Hwa Tan December 3, 2008.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Exact Differentiable Exterior Penalty for Linear Programming Olvi Mangasarian UW Madison & UCSD La Jolla Edward Wild UW Madison December 20, 2015 TexPoint.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Chapter 4 Sensitivity Analysis, Duality and Interior Point Methods.
Inexact SQP methods for equality constrained optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Nonlinear Programming In this handout Gradient Search for Multivariable Unconstrained Optimization KKT Conditions for Optimality of Constrained Optimization.
Optimization in Engineering Design 1 Introduction to Non-Linear Optimization.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Massive Support Vector Regression (via Row and Column Chunking) David R. Musicant and O.L. Mangasarian NIPS 99 Workshop on Learning With Support Vectors.
D Nagesh Kumar, IISc Water Resources Systems Planning and Management: M2L2 Introduction to Optimization (ii) Constrained and Unconstrained Optimization.
1 Chapter 5 Branch-and-bound Framework and Its Applications.
1 Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 23, 2010 Piotr Mirowski Based on slides by Sumit.
Exact Differentiable Exterior Penalty for Linear Programming
Bounded Nonlinear Optimization to Fit a Model of Acoustic Foams
Computational Optimization
CS5321 Numerical Optimization
Chapter 6. Large Scale Optimization
Chap 3. The simplex method
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Computer Animation Algorithms and Techniques
Chapter 6. Large Scale Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Constraints.
Presentation transcript:

Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal January 31, 2007 Inexact Methods for PDE-Constrained Optimization University of North Carolina at Chapel Hill

Nonlinear Optimization “One” problem

Circuit Tuning Building blocks:  Transistors (switches) and Gates (logic units) Improve aspects of the circuit – speed, area, power – by choosing transistor widths AT1 AT3 AT2 d1 d2 w1w2 (A. Wächter, C. Visweswariah, and A. R. Conn, 2005)

Circuit Tuning Building blocks:  Transistors (switches) and Gates (logic units) Improve aspects of the circuit – speed, area, power – by choosing transistor widths Formulate an optimization problem AT1 AT3 AT2 d1 d2 w1w2 (A. Wächter, C. Visweswariah, and A. R. Conn, 2005)

Strategic Bidding Electricity production companies “bid” on how much they will charge for one unit of electricity Independent operator collects bids and sets production schedule and “spot price” to minimize cost to consumers (Pereira, Granville, Dix, and Barroso, 2004)

Strategic Bidding Electricity production companies “bid” on how much they will charge for one unit of electricity Independent operator collects bids and sets production schedule and “spot price” to minimize cost to consumers Bilevel problem Equivalent to MPCC Hard geometry! (Pereira, Granville, Dix, and Barroso, 2004)

Challenges for NLP algorithms Very large problems Numerical noise Availability of derivatives Degeneracies Difficult geometries Expensive function evaluations Real-time solutions needed Integer variables Negative curvature

Outline Problem Formulation  Equality constrained optimization  Sequential Quadratic Programming Inexact Framework  Unconstrained optimization and nonlinear equations  Stopping conditions for linear solver Global Behavior  Merit function and sufficient decrease  Satisfying first order conditions Numerical Results  Model inverse problem  Accuracy tradeoffs Final Remarks  Future work  Negative curvature

Outline Problem Formulation  Equality constrained optimization  Sequential Quadratic Programming Inexact Framework  Unconstrained optimization and nonlinear equations  Stopping conditions for linear solver Global Behavior  Merit function and sufficient decrease  Satisfying first order conditions Numerical Results  Model inverse problem  Accuracy tradeoffs Final Remarks  Future work  Negative curvature

Equality constrained optimization e.g., minimize the difference between observed and expected behavior, subject to atmospheric flow equations (Navier-Stokes) Goal: solve the problem

Equality constrained optimization Define: the Lagrangian Define: the derivatives Goal: solve KKT conditions

Sequential Quadratic Programming (SQP) Algorithm: Newton’s methodAlgorithm: the SQP subproblem Two “equivalent” step computation techniques

Sequential Quadratic Programming (SQP) Algorithm: Newton’s methodAlgorithm: the SQP subproblem Two “equivalent” step computation techniques KKT matrix Cannot be formed Cannot be factored

Sequential Quadratic Programming (SQP) Algorithm: Newton’s methodAlgorithm: the SQP subproblem Two “equivalent” step computation techniques KKT matrix Cannot be formed Cannot be factored Linear system solve Iterative method Inexactness

Outline Problem Formulation  Equality constrained optimization  Sequential Quadratic Programming Inexact Framework  Unconstrained optimization and nonlinear equations  Stopping conditions for linear solver Global Behavior  Merit function and sufficient decrease  Satisfying first order conditions Numerical Results  Model inverse problem  Accuracy tradeoffs Final Remarks  Future work  Negative curvature

Unconstrained optimization Goal: minimize a nonlinear objective Algorithm: Newton’s method (CG)

Unconstrained optimization Goal: minimize a nonlinear objective Algorithm: Newton’s method (CG) Note: choosing any intermediate step ensures global convergence to a local solution of NLP (Steihaug, 1983)

Nonlinear equations Goal: solve a nonlinear system Algorithm: Newton’s method

any step with and ensures descent Nonlinear equations Goal: solve a nonlinear system Algorithm: Newton’s method (Eisenstat and Walker, 1994) (Dembo, Eisenstat, and Steihaug, 1982)

Line Search SQP Framework Define “exact” penalty function Implement a line search

Exact Case

Exact step minimizes the objective on the linearized constraints

Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective

Quadratic/linear model of merit function Create model Quantify reduction obtained from step

Quadratic/linear model of merit function Create model Quantify reduction obtained from step

Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective

Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective … but this is ok since we can account for this conflict by increasing the penalty parameter

Exact Case Exact step minimizes the objective on the linearized constraints … which may lead to an increase in the model objective … but this is ok since we can account for this conflict by increasing the penalty parameter

for k = 0, 1, 2, …  Compute step by…  Set penalty parameter to ensure descent on…  Perform backtracking line search to satisfy…  Update iterate Algorithm Outline (exact steps)

First attempt Proposition: sufficiently small residual 1e-81e-71e-61e-51e-41e-31e-21e-1 Success100% 97% 90%85%72%38% Failure0% 3% 10%15%28%62% Test: 61 problems from CUTEr test set

First attempt… not robust Proposition: sufficiently small residual … not enough for complete robustness  We have multiple goals (feasibility and optimality)  Lagrange multipliers may be completely off … may not have descent!

Recall the line search condition Second attempt Step computation: inexact SQP step We can show

Recall the line search condition Second attempt Step computation: inexact SQP step We can show... but how negative should this be?

for k = 0, 1, 2, …  Compute step  Set penalty parameter to ensure descent  Perform backtracking line search  Update iterate Algorithm Outline (exact steps)

for k = 0, 1, 2, …  Compute step and set penalty parameter to ensure descent and a stable algorithm  Perform backtracking line search  Update iterate Algorithm Outline (inexact steps)

Inexact Case

Step is acceptable if for

Inexact Case Step is acceptable if for

Inexact Case Step is acceptable if for

for k = 0, 1, 2, …  Iteratively solve  Until  Update penalty parameter  Perform backtracking line search  Update iterate Algorithm Outline or

Observe KKT conditions Termination Test

Outline Problem Formulation  Equality constrained optimization  Sequential Quadratic Programming Inexact Framework  Unconstrained optimization and nonlinear equations  Stopping conditions for linear solver Global Behavior  Merit function and sufficient decrease  Satisfying first order conditions Numerical Results  Model inverse problem  Accuracy tradeoffs Final Remarks  Future work  Negative curvature

The sequence of iterates is contained in a convex set and the following conditions hold:  the objective and constraint functions and their first and second derivatives are bounded  the multiplier estimates are bounded  the constraint Jacobians have full row rank and their smallest singular values are bounded below by a positive constant  the Hessian of the Lagrangian is positive definite with smallest eigenvalue bounded below by a positive constant Assumptions

Sufficient Reduction to Sufficient Decrease Taylor expansion of merit function yields Accepted step satisfies

Intermediate Results is bounded below by a positive constant is bounded above

Sufficient Decrease in Merit Function

Step in Dual Space (for sufficiently small and ) Therefore, We converge to an optimal primal solution, and

Outline Problem Formulation  Equality constrained optimization  Sequential Quadratic Programming Inexact Framework  Unconstrained optimization and nonlinear equations  Stopping conditions for linear solver Global Behavior  Merit function and sufficient decrease  Satisfying first order conditions Numerical Results  Model inverse problem  Accuracy tradeoffs Final Remarks  Future work  Negative curvature

Problem Formulation Tikhonov-style regularized inverse problem  Want to solve for a reasonably large mesh size  Want to solve for small regularization parameter SymQMR for linear system solves Input parameters: orRecall: (Curtis and Haber, 2007)

Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res s e s e s e-3 n1024 m512 1e-6 (Curtis and Haber, 2007)

Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res s e s e s e-3 n1024 m512 1e-6 (Curtis and Haber, 2007)

Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 1e s e-2 1e s e-2 1e s e-2 1e s e-2 1e s e-2 n1024 m512 1e-1 (Curtis and Haber, 2007)

Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 1e s e-2 1e s e-2 1e s e-2 1e s e-2 1e s e-2 n8192 m4096 1e-1 (Curtis and Haber, 2007)

Numerical Results Iters.TimeTotal LS Iters. Avg. LS Iters. Avg. Rel. Res. 1e s e-2 1e s e-2 1e s e-2 1e s e-2 1e s e-2 n65536 m e-1 (Curtis and Haber, 2007)

Outline Problem Formulation  Equality constrained optimization  Sequential Quadratic Programming Inexact Framework  Unconstrained optimization and nonlinear equations  Stopping conditions for linear solver Global Behavior  Merit function and sufficient decrease  Satisfying first order conditions Numerical Results  Model inverse problem  Accuracy tradeoffs Final Remarks  Future work  Negative curvature

Review and Future Challenges Review  Defined a globally convergent inexact SQP algorithm  Require only inexact solutions of primal-dual system  Require only matrix-vector products involving objective and constraint function derivatives  Results also apply when only reduced Hessian of Lagrangian is assumed to be positive definite  Numerical experience on model problem is promising Future challenges  (Nearly) Singular constraint Jacobians  Inexact derivative information  Negative curvature  etc., etc., etc….

Negative Curvature Big question  What is the best way to handle negative curvature (i.e., when the reduced Hessian may be indefinite)? Small question  What is the best way to handle negative curvature in the context of our inexact SQP algorithm?  We have no inertia information! Smaller question  When can we handle negative curvature in the context of our inexact SQP algorithm with NO algorithmic modifications?  When do we know that a given step is OK?  Our analysis of the inexact case leads to a few observations…

Why Quadratic Models?

Provides a good… direction? Yes step length? Yes Provides a good… direction? Maybe step length? Maybe

Why Quadratic Models? One can use our stopping criteria as a mechanism for determining which are good directions All that needs to be determined is whether the step lengths are acceptable

Unconstrained Optimization Direct method is the angle test Indirect method is to check the conditions or

Unconstrained Optimization Direct method is the angle test Indirect method is to check the conditions or step qualitystep length

Constrained Optimization Step quality determined by Step length determined by or

Thanks!

Actual Stopping Criteria Stopping conditions: Model reduction condition or

Constraint Feasible Case If feasible, conditions reduce to

Constraint Feasible Case If feasible, conditions reduce to

Constraint Feasible Case If feasible, conditions reduce to Some region around the exact solution

Constraint Feasible Case If feasible, conditions reduce to Ellipse distorted toward the linearized constraints

Constraint Feasible Case If feasible, conditions reduce to