1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.

Slides:



Advertisements
Similar presentations
Curved Trajectories towards Local Minimum of a Function Al Jimenez Mathematics Department California Polytechnic State University San Luis Obispo, CA
Advertisements

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. by Lale Yurttas, Texas A&M University Chapter 14.
Optimization.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 One-Dimensional Unconstrained Optimization Chapter.
Optimization : The min and max of a function
Introducción a la Optimización de procesos químicos. Curso 2005/2006 BASIC CONCEPTS IN OPTIMIZATION: PART II: Continuous & Unconstrained Important concepts.
Optimization of thermal processes
Optimization 吳育德.
Optimization Introduction & 1-D Unconstrained Optimization
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
5/20/ Multidimensional Gradient Methods in Optimization Major: All Engineering Majors Authors: Autar Kaw, Ali.
Steepest Decent and Conjugate Gradients (CG). Solving of the linear equation system.
Performance Optimization
Numerical Optimization
Function Optimization Newton’s Method. Conjugate Gradients
1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me .
Tutorial 12 Unconstrained optimization Conjugate gradients.
Design Optimization School of Engineering University of Bradford 1 Numerical optimization techniques Unconstrained multi-parameter optimization techniques.
Optimization Methods One-Dimensional Unconstrained Optimization
Revision.
Optimization Mechanics of the Simplex Method
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Optimization Methods One-Dimensional Unconstrained Optimization
Optimization Linear Programming and Simplex Method
Unconstrained Optimization Problem
Lecture 17 Today: Start Chapter 9 Next day: More of Chapter 9.
Example 1 Determine whether the stationary point of the following quadratic functions is a local maxima, local minima or saddle point? A point x* is a.
Function Optimization. Newton’s Method Conjugate Gradients Method
Advanced Topics in Optimization
D Nagesh Kumar, IIScOptimization Methods: M2L3 1 Optimization using Calculus Optimization of Functions of Multiple Variables: Unconstrained Optimization.
Why Function Optimization ?
Math for CSLecture 51 Function Optimization. Math for CSLecture 52 There are three main reasons why most problems in robotics, vision, and arguably every.
Optimization Methods One-Dimensional Unconstrained Optimization
Tier I: Mathematical Methods of Optimization
Introduction to Optimization (Part 1)

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
UNCONSTRAINED MULTIVARIABLE
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 9. Optimization problems.
ENCI 303 Lecture PS-19 Optimization 2
84 b Unidimensional Search Methods Most algorithms for unconstrained and constrained optimisation use an efficient unidimensional optimisation technique.
Nonlinear programming Unconstrained optimization techniques.
Chapter 7 Optimization. Content Introduction One dimensional unconstrained Multidimensional unconstrained Example.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Multivariate Unconstrained Optimisation First we consider algorithms for functions for which derivatives are not available. Could try to extend direct.
559 Fish 559; Lecture 5 Non-linear Minimization. 559 Introduction Non-linear minimization (or optimization) is the numerical technique that is used by.
Lecture 13. Geometry Optimization References Computational chemistry: Introduction to the theory and applications of molecular and quantum mechanics, E.
Chapter 10 Minimization or Maximization of Functions.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
1 Chapter 6 General Strategy for Gradient methods (1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Steepest Descent.
Exam 1 Oct 3, closed book Place ITE 119, Time:12:30-1:45pm
Gradient Methods In Optimization
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Exam 1 Oct 3, closed book Place ITE 119, Time:12:30-1:45pm One double-sided cheat sheet (8.5in x 11in) allowed Bring your calculator to the exam Chapters.
Numerical Analysis – Data Fitting Hanyang University Jong-Il Park.
1 Optimization Linear Programming and Simplex Method.
D Nagesh Kumar, IISc Water Resources Systems Planning and Management: M2L2 Introduction to Optimization (ii) Constrained and Unconstrained Optimization.
Non-linear Minimization
Chapter 14.
CS5321 Numerical Optimization
Collaborative Filtering Matrix Factorization Approach
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
Chapter 7 Optimization.
Optimization Part II G.Anuradha.
Outline Unconstrained Optimization Functions of One Variable
Performance Optimization
Outline Preface Fundamentals of Optimization
Section 3: Second Order Methods
Presentation transcript:

1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods

2 Optimization Methods One-Dimensional Unconstrained Optimization Golden-Section Search Quadratic Interpolation Newton's Method Multi-Dimensional Unconstrained Optimization Non-gradient or direct methods Gradient methods Linear Programming (Constrained) Graphical Solution Simplex Method

3 Gradient The gradient vector of a function f, denoted as  f, tells us that from an arbitrary point –Which direction is the steepest ascend/descend? i.e. Direction that will yield the greatest change in f –How much we will gain by taking that step? Indicate by the magnitude of  f = ||  f || 2

4 Gradient – Example Problem: Employ gradient to evaluate the steepest ascent direction for the function f(x, y) = xy 2 at point (2, 2). Solution: 4 unit 8 unit

5 The direction of steepest ascent (gradient) is generally perpendicular, or orthogonal, to the elevation contour.

6 Testing Optimum Point For 1-D problems If f'(x') = 0 and If f"(x') < 0, then x' is a maximum point If f"(x') > 0, then x' is a minimum point If f"(x') = 0, then x' is a saddle point What about for multi-dimensional problems?

7 Testing Optimum Point For 2-D problems, if a point is an optimum point, then In addition, if the point is a maximum point, then Question: If both of these conditions are satisfied for a point, can we conclude that the point is a maximum point?

8 When viewed along the x and y directions. When viewed along the y = x direction. Testing Optimum Point (a, b) is a saddle point

9 For 2-D functions, we also have to take into consideration of That is, whether a maximum or a minimum occurs involves both partial derivatives w.r.t. x and y and the second partials w.r.t. x and y. Testing Optimum Point

10 Also known as the matrix of second partial derivatives. It provides a way to discern if a function has reached an optimum or not. Hessian Matrix (or Hessian of f ) n=2

11 Suppose  f and H is evaluated at x* = (x* 1, x* 2, …, x* n ). If  f = 0, –If H is positive definite, then x* is a minimum point. –If - H is positive definite (or H is negative definite), then x* is a maximum point. –If H is indefinite (neither positive nor negative definite), then x* is a saddle point. –If H is singular, no conclusion (need further investigation) Note: A matrix A is positive definite iff x T Ax > 0 for all non-zero x. A matrix A is positive definite iff the determinants of all its upper left corner sub-matrices are positive. A matrix A is negative definite iff -A is positive definite. Testing Optimum Point (General Case)

12 Assuming that the partial derivatives are continuous at and near the point being evaluated. For function with two variables (i.e. n = 2 ), The quantity |H| is equal to the determinant of the Hessian matrix of f. Testing Optimum Point (Special case – function with two variables)

13 Finite Difference Approximation using Centered-difference approach Used when evaluating partial derivatives is inconvenient.

14 Steepest Ascent Method Steepest ascent method converges linearly. Steepest Ascent Algorithm Select an initial point, x 0 = ( x 1, x 2, …, x n ) for i = 0 to Max_Iteration S i =  f at x i Find h such that f ( x i + hS i ) is maximized x i+1 = x i + hS i Stop loop if x converges or if the error is small enough

15 Example: Suppose f(x, y) = 2xy + 2x – x 2 – 2y 2 Using the steepest ascent method to find the next point if we are moving from point (-1, 1). Next step is to find h that maximize g(h)

16 If h = 0.2 maximizes g(h), then x = -1+6(0.2) = 0.2 and y = 1-6(0.2) = -0.2 would maximize f(x, y). So moving along the direction of gradient from point (-1, 1), we would reach the optimum point (which is our next point) at (0.2, -0.2).

17 Newton's Method One-dimensional Optimization Multi-dimensional Optimization At the optimal Newton's Method H i is the Hessian matrix (or matrix of 2 nd partial derivatives) of f evaluated at x i.

18 Newton's Method Converge quadratically May diverge if the starting point is not close enough to the optimum point. Costly to evaluate H -1

19 Conjugate Direction Methods Conjugate direction methods can be regarded as somewhat in between steepest descent and Newton's method, having the positive features of both of them. Motivation: Desire to accelerate slow convergence of steepest descent, but avoid expensive evaluation, storage, and inversion of Hessian.

20 Conjugate Gradient Approaches (Fletcher-Reeves) ** Methods moving in conjugate directions converge quadratically. Idea: Calculate conjugate direction at each points based on the gradient as Converge faster than Powell's method. Ref: Engineering Optimization (Theory & Practice), 3 rd ed, by Singiresu S. Rao.

21 Marquardt Method ** Idea When a guessed point is far away from the optimum point, use the Steepest Ascend method. As the guessed point is getting closer and closer to the optimum point, gradually switch to the Newton's method.

22 Marquardt Method ** The Marquardt method achieves the objective by modifying the Hessian matrix H in the Newton's Method in the following way: Initially, set α 0 a huge number. Decrease the value of α i in each iteration. When x i is close to the optimum point, makes α i zero (or close to zero).

23 Marquardt Method ** Whenα i is large Whenα i is close to zero Steepest Ascend Method: (i.e., Move in the direction of the gradient.) Newton's Method

24 Summary Gradient – What it is and how to derive Hessian Matrix – What it is and how to derive How to test if a point is maximum, minimum, or saddle point Steepest Ascent Method vs. Conjugate- Gradient Approach vs. Newton Method