Chapter 10 Minimization or Maximization of Functions.

Slides:



Advertisements
Similar presentations
Optimization.
Advertisements

Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 One-Dimensional Unconstrained Optimization Chapter.
Optimization : The min and max of a function
Optimization of thermal processes
Optimization 吳育德.
Optimization methods Review
Optimization Introduction & 1-D Unconstrained Optimization
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Direction Set (Powell’s) Methods in Multidimensions
Steepest Decent and Conjugate Gradients (CG). Solving of the linear equation system.
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
1cs542g-term Notes  Assignment 1 due tonight ( me by tomorrow morning)
Numerical Optimization
Function Optimization Newton’s Method. Conjugate Gradients
Tutorial 12 Unconstrained optimization Conjugate gradients.
Engineering Optimization
Optimization Methods One-Dimensional Unconstrained Optimization
Gradient Methods May Preview Background Steepest Descent Conjugate Gradient.
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Optimization Methods One-Dimensional Unconstrained Optimization
Unconstrained Optimization Problem
Gradient Methods Yaron Lipman May Preview Background Steepest Descent Conjugate Gradient.
Function Optimization. Newton’s Method Conjugate Gradients Method
Advanced Topics in Optimization
Maximum and Minimum Values (Section 3.1)
Why Function Optimization ?
Optimization Methods One-Dimensional Unconstrained Optimization
Molecular Modeling: Geometry Optimization C372 Introduction to Cheminformatics II Kelsey Forsythe.
Ch. 11: Optimization and Search Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 some slides from Stephen Marsland, some images.

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Computational Optimization
Molecular Modeling: Geometry Optimization C372 Introduction to Cheminformatics II Kelsey Forsythe.
UNCONSTRAINED MULTIVARIABLE
Simulation of Random Walk How do we investigate this numerically? Choose the step length to be a=1 Use a computer to generate random numbers r i uniformly.
Geometry Optimisation Modelling OH + C 2 H 4 *CH 2 -CH 2 -OH CH 3 -CH 2 -O* 3D PES.
Chapter 17 Boundary Value Problems. Standard Form of Two-Point Boundary Value Problem In total, there are n 1 +n 2 =N boundary conditions.
ENCI 303 Lecture PS-19 Optimization 2
84 b Unidimensional Search Methods Most algorithms for unconstrained and constrained optimisation use an efficient unidimensional optimisation technique.
Nonlinear programming Unconstrained optimization techniques.
Chapter 7 Optimization. Content Introduction One dimensional unconstrained Multidimensional unconstrained Example.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
Computer Animation Rick Parent Computer Animation Algorithms and Techniques Optimization & Constraints Add mention of global techiques Add mention of calculus.
Multivariate Unconstrained Optimisation First we consider algorithms for functions for which derivatives are not available. Could try to extend direct.
Case Study in Computational Science & Engineering - Lecture 5 1 Iterative Solution of Linear Systems Jacobi Method while not converged do { }
Quasi-Newton Methods of Optimization Lecture 2. General Algorithm n A Baseline Scenario Algorithm U (Model algorithm for n- dimensional unconstrained.
Lecture 13. Geometry Optimization References Computational chemistry: Introduction to the theory and applications of molecular and quantum mechanics, E.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
1 Chapter 6 General Strategy for Gradient methods (1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Steepest Descent.
Gradient Methods In Optimization
Signal & Weight Vector Spaces
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Advanced Computer Graphics Optimization Part 2 Spring 2002 Professor Brogan.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Section 15.3 Constrained Optimization: Lagrange Multipliers.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Computational Biology BS123A/MB223 UC-Irvine Ray Luo, MBB, BS.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 2 - Chapter 7 Optimization.
Announcements Topics: -sections (differentiation rules), 5.6, and 5.7 * Read these sections and study solved examples in your textbook! Work On:
Chapter 10. Numerical Solutions of Nonlinear Systems of Equations
Chapter 7 Optimization.
~ Least Squares example
Chapter 10 Minimization or Maximization of Functions
~ Least Squares example
Part 4 - Chapter 13.
Performance Optimization
Section 3: Second Order Methods
Presentation transcript:

Chapter 10 Minimization or Maximization of Functions

Optimization Problems Solution of equations can be formulated as an optimization problem, e.g., density functional theory in electronic structure, conformation of proteins, etc Minimization with constraints – operations research (linear programming, optimal conditions in management science, traveling salesman problem, etc)

General Consideration Use function values only, or use function values and its derivatives Storage of O(N) or O(N 2 ) With constraints or no constraints Choice of methods

Local & Global Extremum

Bracketing and Search in 1D Bracket a minimum means that for given a < b < c, we have f(b) < f(a), and f(b) < f(c). There is a minimum in the interval (a,c). a b c

How accurate can we locate a minimum? Let b a minimum of function f(x),Taylor expanding around b, we have The best we can do is when the second correction term reaches machine epsilon comparing to the function value, so

Golden Section Search Choose x such that the ratio of intervals [a,b] to [b,c] is the same as [a,x] to [x,b]. Remove [a,x] if f[x] > f[b], or remove [b,c] if f[x] < f[b]. The asymptotic limit of the ratio is the Golden mean ab c x

Parabolic Interpolation & Brent’s Method Brent’s method combines parabolic interpolation with Golden section search, with some complicated bookkeeping. See NR, page for details.

Strategy in Higher Dimensions 1.Starting from a point P and a direction n, find the minimum on the line P + n, i.e., do a 1D minimization of y( )=f(P+ n) 2.Replace P by P + min n, choose another direction n’ and repeat step 1. The trick and variation of the algorithms are on chosen n.

Local Properties near Minimum Let P be some point of interest which is at the origin x=0. Taylor expansion gives Minimizing f is the same as solving the equation T for transpose of a matrix

Search along Coordinate Directions Search minimum along x direction, followed by search minimum along y direction, and so on. Such method takes a very large number of steps to converge. The curved loops represent f(x,y) = const.

Steepest Descent Search in the direction with the largest decrease, i.e., n = -  f Constant f contour line (surface) is perpendicular to n, because df = dx  f = 0. The current search direction n and next search direction are orthogonal, because for minimum we have y’( ) = df(P+ n)/d = n T   f| P+ n = 0 n n’ n T  n’ = 0

Conjugate Condition n 1 T  A  n 2 = 0 Make a linear coordinate transformation, such that contour is circular and (search) vectors are orthogonal

Conjugate Gradient Method 1.Start with steepest descent direction n 0 = g 0 = -  f(x 0 ), find new minimum x 1 2.Build the next search direction n 1 from g 0 and g 1 = -  f(x 1 ), such that n 0  A  n 1 = 0 3.Repeat step 2 iteratively to find n j (a Gram-Schmidt orthogonalization process). The result is a set of N vectors (in N dimensions) n i T  A  n j = 0

Conjugate Gradient Algorithm 1.Initialize n 0 = g 0 = -  f(x 0 ), i = 0, 2.Find that minimizes f(x i + n i ), let x i+1 =x i + n i 3.Compute new negative gradient g i+1 = -  f(x i+1 ) 4.Compute 5.Update new search direction as n i+1 = g i+1 +  i n i ; ++ i, go to 2 (Fletcher-Reeves)

The Conjugate Gradient Program

Simulated Annealing To minimize f(x), we make random change to x by the following rule: Set T a large value, decrease as we go Metropolis algorithm: make local change from x to x’. If f decreases, accept the change, otherwise, accept only with a small probability r = exp [ - ( f(x’)-f(x) ) /T ]. This is done by comparing r with a random number 0 < ξ < 1.

Traveling Salesman Problem Singapore Kuala Lumpur Hong Kong Taipei Shanghai Beijing Tokyo Find shortest path that cycles through each city exactly once.

Problem set 7 1.Suppose that the function is given by the quadratic form f=(1/2)x T  A  x, where A is a symmetric and positive definite matrix. Find a linear transform to x so that in the new coordinate system, the function becomes f = (1/2)|y| 2, y = Ux [i.e., the contour is exactly circular or spherical]. If two vectors in the new system are orthogonal, y 1 T  y 2 =0, what does it mean in the original system? 2.We’ll discuss the conjugate gradient method in some more detail following the paper: conjugate-gradient.pdf conjugate-gradient.pdf