Presentation is loading. Please wait.

Presentation is loading. Please wait.

Gradient Methods April 2004. Preview Background Steepest Descent Conjugate Gradient.

Similar presentations


Presentation on theme: "Gradient Methods April 2004. Preview Background Steepest Descent Conjugate Gradient."— Presentation transcript:

1 Gradient Methods April 2004

2 Preview Background Steepest Descent Conjugate Gradient

3 Preview Background Steepest Descent Conjugate Gradient

4 Background Motivation The gradient notion The Wolfe Theorems

5 Motivation The min(max) problem: But we learned in calculus how to solve that kind of question!

6 Motivation Not exactly, Functions: High order polynomials: What about function that don ’ t have an analytic presentation: “ Black Box ”

7 Motivation- “ real world ” problem Connectivity shapes (isenburg,gumhold,gotsman) What do we get only from C without geometry?

8 Motivation- “ real world ” problem First we introduce error functionals and then try to minimize them:

9 Motivation- “ real world ” problem Then we minimize: High dimension non-linear problem. The authors use conjugate gradient method which is maybe the most popular optimization technique based on what we ’ ll see here.

10 Motivation- “ real world ” problem Changing the parameter:

11 Motivation General problem: find global min(max) This lecture will concentrate on finding local minimum.

12 Background Motivation The gradient notion The Wolfe Theorems

13

14 Directional Derivatives: first, the one dimension derivative:

15 Directional Derivatives : Along the Axes …

16 Directional Derivatives : In general direction …

17 Directional Derivatives

18 In the plane The Gradient: Definition in

19 The Gradient: Definition

20 The Gradient Properties The gradient defines (hyper) plane approximating the function infinitesimally

21 The Gradient properties By the chain rule: (important for later use)

22 The Gradient properties Proposition 1: is maximal choosing is minimal choosing (intuitive: the gradient point the greatest change direction)

23 The Gradient properties Proof: (only for minimum case) Assign: by chain rule:

24 The Gradient properties On the other hand for general v:

25 The Gradient Properties Proposition 2: let be a smooth function around P, if f has local minimum (maximum) at p then, (Intuitive: necessary for local min(max))

26 The Gradient Properties Proof: Intuitive:

27 The Gradient Properties Formally: for any We get:

28 The Gradient Properties We found the best INFINITESIMAL DIRECTION at each point, Looking for minimum: “ blind man ” procedure How can we derive the way to the minimum using this knowledge?

29 Background Motivation The gradient notion The Wolfe Theorems

30 The Wolfe Theorem This is the link from the previous gradient properties to the constructive algorithm. The problem:

31 The Wolfe Theorem We introduce a model for algorithm: Data: Step 0:set i=0 Step 1:ifstop, else, compute search direction Step 2: compute the step-size Step 3:setgo to step 1

32 The Wolfe Theorem The Theorem: suppose C1 smooth, and exist continuous function: And, And, the search vectors constructed by the model algorithm satisfy:

33 The Wolfe Theorem And Then if is the sequence constructed by the algorithm model, then any accumulation point y of this sequence satisfy:

34 The Wolfe Theorem The theorem has very intuitive interpretation : Always go in decent direction.

35 Preview Background Steepest Descent Conjugate Gradient

36 Steepest Descent What it mean? We now use what we have learned to implement the most basic minimization technique. First we introduce the algorithm, which is a version of the model algorithm. The problem:

37 Steepest Descent Steepest descent algorithm: Data: Step 0:set i=0 Step 1:ifstop, else, compute search direction Step 2: compute the step-size Step 3:setgo to step 1

38 Steepest Descent Theorem: if is a sequence constructed by the SD algorithm, then every accumulation point y of the sequence satisfy: Proof: from Wolfe theorem Remark: wolfe theorem gives us numerical stability is the derivatives aren ’ t given (are calculated numerically).

39 Steepest Descent From the chain rule: Therefore the method of steepest descent looks like this:

40 Steepest Descent

41 The steepest descent find critical point and local minimum. Implicit step-size rule Actually we reduced the problem to finding minimum: There are extensions that gives the step size rule in discrete sense. (Armijo)

42 Steepest Descent Back with our connectivity shapes: the authors solve the 1-dimension problem analytically. They change the spring energy and get a quartic polynomial in x

43 Preview Background Steepest Descent Conjugate Gradient


Download ppt "Gradient Methods April 2004. Preview Background Steepest Descent Conjugate Gradient."

Similar presentations


Ads by Google