Presentation is loading. Please wait.

Presentation is loading. Please wait.

UNCONSTRAINED MULTIVARIABLE

Similar presentations


Presentation on theme: "UNCONSTRAINED MULTIVARIABLE"— Presentation transcript:

1 UNCONSTRAINED MULTIVARIABLE
Chapter 6 UNCONSTRAINED MULTIVARIABLE OPTIMIZATION Chapter 6

2 Chapter 6 6.1 Function Values Only
First Derivatives of f (gradient and conjugate direction methods) Second Derivatives of f (e.g., Newton’s method) Quasi-Newton methods Chapter 6

3 Chapter 6

4 Chapter 6

5 Chapter 6

6 Chapter 6

7 Chapter 6

8 Chapter 6

9 General Strategy for Gradient methods
(1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Chapter 6 Steepest Descent Search Direction Don’t need to normalize Method terminates at any stationary point. Why?

10 Chapter 6 So procedure can stop at saddle point. Need to show
is positive definite for a minimum. Step Length How to pick a analytically numerically Chapter 6

11 Chapter 6

12 Chapter 6

13 Chapter 6

14 Chapter 6 Analytical Method
How does one minimize a function in a search direction using an analytical method? It means s is fixed and you want to pick a, the step length to minimize f(x). Note Chapter 6 (6.9) This yields a minimum of the approximating function.

15 Chapter 6 Numerical Method Use coarse search first
(1) Fixed a (a = 1) or variable a (a = 1, 2, ½, etc.) Options for optimizing a (1) Use interpolation such as quadratic, cubic (2) Region Elimination (Golden Search) (3) Newton, Secant, Quasi-Newton (4) Random (5) Analytical optimization (1), (3), and (5) are preferred. However, it may not be desirable to exactly optimize a (better to generate new search directions). Chapter 6

16 Suppose we calculate the gradient at the point xT = [2 2]
Chapter 6

17 Chapter 6

18 Chapter 6

19 Chapter 6 Termination Criteria f(x)
Big change in f(x) but little change in x. Code will stop if Dx is sole criterion. x f(x) Big change in x but little change in f(x). Code will stop if Dx is sole criterion. Chapter 6 x For minimization you can use up to three criteria for termination: (1) (2) (3)

20 Conjugate Search Directions
Improvement over gradient method for general quadratic functions Basis for many NLP techniques Two search directions are conjugate relative to Q if To minimize f(xnx1) when H is a constant matrix (=Q), you are guaranteed to reach the optimum in n conjugate direction stages if you minimize exactly at each stage (one-dimensional search) Chapter 6

21 Chapter 6

22 Conjugate Gradient Method
by minimizing f(x) with respect to a in the s0 direction (i.e., carry out a unidimensional search for a0). Step 3. Calculate The new search direction is a linear combination of Chapter 6 For the kth iteration the relation is (6.6) For a quadratic function it can be shown that these successive search directions are conjugate. After n iterations (k = n), the quadratic function is minimized. For a nonquadratic function, the procedure cycles again with xn+1 becoming x0. Step 4. Test for convergence to the minimum of f(x). If convergence is not attained, return to step 3. Step n. Terminate the algorithm when is less than some prescribed tolerance.

23 Chapter 6

24 Chapter 6

25 Chapter 6

26 Chapter 6

27 Chapter 6 Chapter 6 Minimize
using the method of conjugate gradients with as an initial point. In vector notation, For steepest descent, Chapter 6 Chapter 6 Steepest Descent Step (1-D Search) The objective function can be expressed as a function of a0 as follows: Minimizing f(a0), we obtain f = at a0 = Hence

28 Chapter 6 Chapter 6 Calculate Weighting of Previous step
The new gradient can now be determined as and w 0 can be computed as Generate New (Conjugate) Search Direction Chapter 6 Chapter 6 and One dimensional Search Solving for a1 as before [i.e., expressing f(x1) as a function of a1 and minimizing with respect to a1] yields f = 5.91 x at a1 = Hence which is the optimum (in 2 steps, which agrees with the theory).

29 Chapter 6

30 Chapter 6

31 Chapter 6

32 Fletcher – Reeves Conjugate Gradient Method
Chapter 6 Derivation:

33 Chapter 6 and solving for the weighting factor:

34 Linear vs. Quadratic Approximation of f(x)
Chapter 6

35 Chapter 6

36 Chapter 6

37 Chapter 6

38 Chapter 6

39 Marquardt’s Method Chapter 6

40 Step 1 Step 2 Chapter 6 Step 3 Step 4

41 Step 5 Step 6 Chapter 6 Step 7 Step 8 Step 9

42 Chapter 6 Secant Methods
Recall for one dimensional search the secant method only uses values of f(x) and f ′(x). Chapter 6

43 Chapter 6

44 Chapter 6 • Probably the best update formula is the BFGS update
(Broyden – Fletcher – Goldfarb – Shanno) – ca. 1970 • BFGS is the basis for the unconstrained optimizer in the Excel Solver • Does not require inverting the Hessian matrix but approximates the inverse with values of Chapter 6

45 Chapter 6

46 Chapter 6

47 Chapter 6

48 Chapter 6


Download ppt "UNCONSTRAINED MULTIVARIABLE"

Similar presentations


Ads by Google