Review + Announcements 2/22/08. Presentation schedule Friday 4/25 (5 max)Tuesday 4/29 (5 max) 1. Miguel Jaller 8:031. Jayanth 8:03 2. Adrienne Peltz 8:202.

Slides:



Advertisements
Similar presentations
Solving LP Models Improving Search Special Form of Improving Search
Advertisements

Optimization.
Engineering Optimization
Introducción a la Optimización de procesos químicos. Curso 2005/2006 BASIC CONCEPTS IN OPTIMIZATION: PART II: Continuous & Unconstrained Important concepts.
Optimization 吳育德.
1 TTK4135 Optimization and control B.Foss Spring semester 2005 TTK4135 Optimization and control Spring semester 2005 Scope - this you shall learn Optimization.
C&O 355 Mathematical Programming Fall 2010 Lecture 15 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
1 OR II GSLM Outline  classical optimization – unconstrained optimization  dimensions of optimization  feasible direction.
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
Easy Optimization Problems, Relaxation, Local Processing for a small subset of variables.
Separating Hyperplanes
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
The Most Important Concept in Optimization (minimization)  A point is said to be an optimal solution of a unconstrained minimization if there exists no.
1cs542g-term Notes  Assignment 1 due tonight ( me by tomorrow morning)
MIT and James Orlin © Nonlinear Programming Theory.
Numerical Optimization
1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me .
Nonlinear Optimization for Optimal Control
Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.
Reformulated - SVR as a Constrained Minimization Problem subject to n+1+2m variables and 2m constrains minimization problem Enlarge the problem size and.
Exploiting Duality (Particularly the dual of SVM) M. Pawan Kumar VISUAL GEOMETRY GROUP.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
Unconstrained Optimization Problem
Advanced Topics in Optimization
Optimality Conditions for Nonlinear Optimization Ashish Goel Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.
Computer Algorithms Mathematical Programming ECE 665 Professor Maciej Ciesielski By DFG.
Tier I: Mathematical Methods of Optimization
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Computational Optimization
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 9. Optimization problems.
KKT Practice and Second Order Conditions from Nash and Sofer
Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal February 12, 2007 Inexact Methods for PDE-Constrained Optimization.
ENCI 303 Lecture PS-19 Optimization 2
Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal January 31, 2007 Inexact Methods for PDE-Constrained Optimization.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Nonlinear Programming Models
Advanced Operations Research Models Instructor: Dr. A. Seifi Teaching Assistant: Golbarg Kazemi 1.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
The Method of Moving Asymptotes
Exact Differentiable Exterior Penalty for Linear Programming Olvi Mangasarian UW Madison & UCSD La Jolla Edward Wild UW Madison December 20, 2015 TexPoint.
Chapter 4 Sensitivity Analysis, Duality and Interior Point Methods.
CPSC 536N Sparse Approximations Winter 2013 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAA.
Inexact SQP methods for equality constrained optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Nonlinear Programming In this handout Gradient Search for Multivariable Unconstrained Optimization KKT Conditions for Optimality of Constrained Optimization.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Linear & Nonlinear Programming -- Basic Properties of Solutions and Algorithms.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
Inequality Constraints Lecture 7. Inequality Contraints (I) n A Review of Lagrange Multipliers –As we discussed last time, the first order necessary conditions.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
OR II GSLM
Regularized Least-Squares and Convex Optimization.
Optimal Control.
Bounded Nonlinear Optimization to Fit a Model of Acoustic Foams
Chapter 11 Optimization with Equality Constraints
CSLT ML Summer Seminar (13)
Computational Optimization
Computational Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
~ Least Squares example
CS5321 Numerical Optimization
I.4 Polyhedral Theory (NW)
~ Least Squares example
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
Constraints.
Presentation transcript:

Review + Announcements 2/22/08

Presentation schedule Friday 4/25 (5 max)Tuesday 4/29 (5 max) 1. Miguel Jaller 8:031. Jayanth 8:03 2. Adrienne Peltz 8:202. Raghav 9;20 3. Olga Grisin 8:37 3. Rhyss 8:37 4. Dan Erceg 8:544. Tim *:54 5. Nick Suhr 9: Lindsey Garret and Mark Yuhas 9:11 6. Christos Boutsidis 9:28 Monday 4/28 4:00 7:00 Pizza included Lisa Pak Christos Boutsidis David Doria. Zhi Zeng Carlos Varun Samrat Matt Adarsh Ramsuhramonian Be on time. Plan your presentation for 15 minutes. Strict schedule. Suggest putting presentation in Your public_html directory in rcs so you can click and go. Monday night class is in Amos Eaton to 7.

Other Dates Project Papers due Friday (or in class Monday if you have a Friday presentation) Final Tuesday 5/6 3 p.m. Eaton 214 Open book/note (no computers) Comprehensive. Labs fair game too. Office hours Monday 5/5 10 to 12 (or )

What did we learn? Theme 1: “There is nothing more practical than a good theory” - Kurt Lewin Algorithm arise out of the optimality conditions.

What did we learn? Theme 2: To solve a harder problem, reduce it to an easier problem that you already know how to solve.

Fundamental Theoretical Ideas Convex functions and sets Convex programs Differentiability Taylor Series Approximations Descent Directions Combining these with the ideas of feasible directions provides the basis for optimality conditions.

Convex Functions A function f is (strictly) convex on a convex set S, if and only if for any x,y  S, f( x+(1- )y)(<)  f(x)+ (1- )f(y) for all 0   1. x y f(y) f(x) λx+(1- )y f(λx+(1- )y)

Convex Sets A set S is convex if the line segment joining any two points in the set is also in the set, i.e., for any x,y  S, x+(1- )y  S for all 0   1 }. convex not convex

Convex Program min f(x) subject to x  S where f and S are convex Make optimization nice Many practical problems are convex problem Use convex program as subproblem for nonconvex programs

Theorem : Global Solution of convex program If x* is a local minimizer of a convex programming problem, x* is also a global minimizer. Further more if the objective is strictly convex then x* is the unique global minimizer. Proof: contradiction x* y f(y)<f(x*)

First Order Taylor Series Approximation Let x=x*+p Says that a linear approximation of a function works well locally f(x) x*

Second Order Taylor Series Approximation Let x=x*+p Says that a quadratic approximation of a function works even better locally x* f(x)

Descent Directions If the directional derivative is negative then linesearch will lead to decrease in the function [8,2] [0,-1] d

First Order Necessary Conditions Theorem: Let f be continuously differentiable. If x* is a local minimizer of (1), then

Second Order Sufficient Conditions Theorem: Let f be twice continuously differentiable. Ifand then x* is a strict local minimizer of (1).

Second Order Necessary Conditions Theorem: Let f be twice continuously differentiable. If x* is a local minimizer of (1) then

Optimality Conditions First Order Necessary Second Order Necessary Second Order Sufficient With convexity the necessary conditions become sufficient.

Easiest Problem Line Search = 1-D Optimization Optimality conditions based on first and second derivatives Golden section search (1)

Sometimes can solve linesearch exactly The exact stepsize can be found

General Optimization algorithm Specify some initial guess x 0 For k = 0, 1, …… If x k is optimal then stop Determine descent direction p k Determine improved estimate of the solution: x k+1 =x k + k p k Last step is one-dimensional search problem called line search

Newton’s Method Minimizing quadratic has closed form

General nonlinear functions For non-quadratic f (twice cont. diff): Approximate by 2 nd order TSA Solve for FONC for quadratic approx.

Basic Newton’s Algorithm Start with x 0 For k =1,…,K If x k is optimal then stop Solve: X k+1 =x k +p

Final Newton’s Algorithm Start with x 0 For k =1,…,K If x k is optimal then stop Solve: using modified cholesky factorization Perform linesearch to determine X k+1 =x k +  k p k What are pros and cons?

Steepest Descent Algorithm Start with x 0 For k =1,…,K If x k is optimal then stop Perform exact or backtracking linesearch to determine x k+1 =x k +  k p k

Inexact linesearch can work quite well too! For 0<c 1 <c 2 <1 Solution exists for any descent direction if f is bounded below on the linesearch. (Lemma 3.1)

Conditioning Important for gradient methods! 50(x-10)^2+y^2 Cond num =50/1=50 Steepest Descent ZIGZAGS!!! Know Pros and Cons of each approach

Conjugate Gradient (CG) Method for minimizing quadratic function Low storage method CG only stores vector information CG superlinear convergence for nice problems or when properly scaled Great for solving QP subproblems

Quasi Newton Methods Pros and Cons Globally converges to a local min always find descent direction Superlinear convergence Requires only first order information – approximates Hessian More complicated than steepest descent Requires sophisticated linear algebra Have to watch out for numerical error

Quasi Newton Methods Pros and Cons Globally converges to a local min Superlinear convergence w/o computing Hessian Works great in practice. Widely used. More complicated than steepest descent Best implementations require sophisticated linear algebra, linesearch, dealing with curvature conditions. Have to watch out for numerical error.

Trust Region Methods Alternative to line search methods Optimize quadratic model of objective within the “trust region”

Easiest Problem Linear equality constraints

Lemma 14.1 Necessary Conditions (Nash + Sofer) If x* is a local min of f over {x|Ax=b}, and Z is a null matrix Or equivalently use KKT Conditions Other conditions Generalize similarly

Handy ways to compute Null Space Variable Reduction Method Orthogonal Projection Matrix QR factorization (best numerically) Z=Null(A) in matlab

Next Easiest Problem Linear equality constraints Constraints form a polyhedron

Polyhedron Ax>=b a 1 x = b 1 a1a1 Inequality Case a 4 x = b 4 a 3 x = b 3 a 5 x = b 5 a 2 x = b 5 a2a2 Inequality problem Nonnegative Multipliers imply gradient points to the greater than Side of the constraint. x* Inequality FONC:

Second Order Sufficient Conditions for Linear Inequalities If (x*, *) satisfies

Sufficient Conditions for Linear Inequalities where Z + is a basis matrix for Null(A + ) and A + corresponds to nondegenerate active constraints) i.e.

General Constraints Careful : Sufficient conditions are the same as before Necessary conditions have extra constraint qualification to make sure Lagrangian multipliers exist!

Necessary Conditions General If x* satisfies LICQ and is a local min of f over {x|g(x)>=0,h(x)=0},

Algorithms build on prior Approaches Linear Equality Constrained: Convert to unconstrained and solve Different ways to represent Null space produce Algorithms in practice

Prior Approaches (cont) Linear Inequality Constrained: Identify active constraints Solve equality constrained subproblems Nonlinear Inequality Constrained: Linearize constraints Solve subproblems

Change one item of working set at a time Active Set Methods NW 16.5

Traverse interior of set (a little more later) Interior point algorithms NW 16.6

Change many elements of working set at once Gradient Projection NW 16.7

Generic inexact penalty problem From To What are penalty problems and why do we use them? Difference between exact and inexact penalties.

Augmented Lagrangian Consider min f(x) s.t h(x)=0 Start with L(x, )=f(x)- ’h(x) Add penalty L(x,,c)=f(x)- ’h(x)+μ/2||h(x)|| 2 The penalty helps insure that the point is feasible. Why do we like these? How do they work in practice?

Sequential Quadratic Programming (SQP) Basic Idea: QP with constraints are easy. For any guess of active constraints, just have to solve system of equations. So why not solve general problem as a series of constrained QPs. Which QP should be used?

Trust Region Works Great We only trust approximation locally so limit step to this region by adding constraint to QP Trust region No stepsize needed!

Advanced topics Duality Theory – Can choose to solve primal or dual problem. Dual is always nice. But there may be a “duality gap” if overall problem is not nice. Nonsmooth optimization Can do the whole thing again on the basis of subgradients instead of gradients.

Subgradient Generalization of the gradient Definition Hinge loss 01