Download presentation

1
**Optimization with Constraints**

Standard formulation Why are constraints a problem? Canyons in design space Equality constraints are particularly bad!

2
**Derivative-based optimizers**

All are predicated on the assumption that function evaluations are expensive and gradients can be calculated. Similar to a person put at night on a hill and directed to find the lowest point in an adjacent valley using a flashlight with limited battery Basic strategy: Flash light to get derivative and select direction. Go straight in that direction until you start going up or hit constraint. Repeat until converged. Steepest descent does not work well. For unconstrained problems use quasi-Newton methods.

3
**Gradient projection and reduced gradient methods**

Find good direction tangent to active constraints Move a distance and then restore to constraint boundaries How do we decide when to stop and restore?

4
**The Feasible Directions method**

Compromise constraint avoidance and objective reduction

5
**Penalty function methods**

Quadratic penalty function Gradual rise of penalty parameter leads to sequence of unconstrained minimization technique (SUMT). Why is it important?

6
Example 5.7.1

7
**Contours for r=1 . >> x=linspace(1,4,40);**

>> y=linspace(0.1,0.5,41); >> [X,Y]=meshgrid(x,y); >> Z=X.^2+10*Y.^2+r*(4-X-Y).^2; >> cs=contour(X,Y,Z); >> clabel(cs);

8
Contours for r=10 .

9
Contours for r=100 .

10
Contours for r=1000 . For non-derivative methods can avoid this by having penalty proportional to absolute value of violation instead of its square!

11
Multiplier methods By adding the Lagrange multipliers to penalty term can avoid ill-conditioning associated with high penalty

12
**Can estimate Lagrange multipliers**

Stationarity Without penalty So: Iterate See example in textbook

13
**5.9: Projected Lagrangian methods**

Sequential quadratic programming Find direction by solving Find alpha by minimizing

14
**Matlab function fmincon**

FMINCON attempts to solve problems of the form: min F(X) subject to: A*X <= B, Aeq*X = Beq (linear cons) X C(X) <= 0, Ceq(X) = 0 (nonlinear cons) LB <= X <= UB X=FMINCON(FUN,X0,A,B,Aeq,Beq,LB,UB,NONLCON) subjects the minimization to the constraints defined in NONLCON. The function NONLCON accepts X and returns the vectors C and Ceq, representing the nonlinear inequalities and equalities respectively. (Set LB=[] and/or UB=[] if no bounds exist.) [X,FVAL]=FMINCON(FUN,X0,...) returns the value of the objective function FUN at the solution X. What does the order of arguments in fmincon tell you about the expectations of Matlab on the most common problems to be solved?

15
**Useful additional information**

[X,FVAL,EXITFLAG] = fmincon(FUN,X0,...) returns an EXITFLAG that describes the exit condition of fmincon. Possible values of EXITFLAG 1 First order optimality conditions satisfied. 0 Too many function evaluations or iterations. -1 Stopped by output/plot function. -2 No feasible point found. [X,FVAL,EXITFLAG,OUTPUT,LAMBDA] = fmincon(FUN,X0,...) returns Lagrange multipliers at the solution X: LAMBDA.lower for LB, LAMBDA.upper for UB, LAMBDA.ineqlin is for the linear inequalities, LAMBDA.eqlin is for the linear equalities, LAMBDA.ineqnonlin is for the nonlinear inequalities, and LAMBDA.eqnonlin is for the nonlinear equalities.

16
Ring Example Quadratic function and constraint function f=quad2(x) Global a f=x(1)^2+a*x(2)^2; function [c,ceq]=ring(x) global ri ro c(1)=ri^2-x(1)^2-x(2)^2; c(2)=x(1)^2+x(2)^2-ro^2; ceq=[]; x0=[1,10]; a=10;ri=10.; ro=20; x = fval =

17
**Making it harder for fmincon**

Warning: Trust-region-reflective method does not currently solve this type of problem, using active-set (line search) instead. > In fmincon at 437 Maximum number of function evaluations exceeded; increase OPTIONS.MaxFunEvals. x = fval =

18
**Restart sometimes helps**

>>x0=x x0 = >> x = fval = Why restarting a solution is often more effective than having a large number of iterations in the original search?

Similar presentations

OK

Survey of gradient based constrained optimization algorithms Select algorithms based on their popularity. Additional details and additional algorithms.

Survey of gradient based constrained optimization algorithms Select algorithms based on their popularity. Additional details and additional algorithms.

© 2018 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on cold call teach like a champion Ppt on dhaka stock exchange Ppt on web server Psychology ppt on personality Mis ppt on nokia cell Ppt on the rise of nationalism in europe class 10 Ppt on corporate social responsibility under companies act 2013 Ppt on online mobile shopping Ppt on product advertising Ppt on edge detection method