1cs542g-term1-2007 Notes  Extra class this Friday 1-2pm  If you want to receive emails about the course (and are auditing) send me email.

Slides:



Advertisements
Similar presentations
Optimization.
Advertisements

Engineering Optimization
Optimization : The min and max of a function
Optimization 吳育德.
Least Squares example There are 3 mountains u,y,z that from one site have been measured as 2474 ft., 3882 ft., and 4834 ft.. But from u, y looks 1422 ft.
Steepest Decent and Conjugate Gradients (CG). Solving of the linear equation system.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
1cs542g-term Notes  Extra class this Friday 1-2pm  Assignment 2 is out  Error in last lecture: quasi-Newton methods based on the secant condition:
Easy Optimization Problems, Relaxation, Local Processing for a single variable.
1Notes  Handing assignment 0 back (at the front of the room)  Read the newsgroup!  Planning to put 16mm films on the web soon (possibly tomorrow)
Jonathan Richard Shewchuk Reading Group Presention By David Cline
1cs542g-term Notes  Assignment 1 due tonight ( me by tomorrow morning)
PHYS2020 NUMERICAL ALGORITHM NOTES ROOTS OF EQUATIONS.
Numerical Optimization
Unconstrained Optimization Rong Jin. Recap  Gradient ascent/descent Simple algorithm, only requires the first order derivative Problem: difficulty in.
458 Interlude (Optimization and other Numerical Methods) Fish 458, Lecture 8.
Tutorial 12 Unconstrained optimization Conjugate gradients.
Methods For Nonlinear Least-Square Problems
Optimization Methods One-Dimensional Unconstrained Optimization
1cs542g-term Notes  Note that r 2 log(r) is NaN at r=0: instead smoothly extend to be 0 at r=0  Schedule a make-up lecture?
Gradient Methods May Preview Background Steepest Descent Conjugate Gradient.
Optimization Methods One-Dimensional Unconstrained Optimization
Unconstrained Optimization Problem
Gradient Methods Yaron Lipman May Preview Background Steepest Descent Conjugate Gradient.
Engineering Optimization
Advanced Topics in Optimization
CSE 245: Computer Aided Circuit Simulation and Verification
Why Function Optimization ?
1cs426-winter-2008 Notes  SIGGRAPH crunch time - my apologies :-)
Optimization Methods One-Dimensional Unconstrained Optimization
Unconstrained Optimization Rong Jin. Logistic Regression The optimization problem is to find weights w and b that maximizes the above log-likelihood How.
Tier I: Mathematical Methods of Optimization

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
Computational Optimization
UNCONSTRAINED MULTIVARIABLE
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 9. Optimization problems.
Collaborative Filtering Matrix Factorization Approach
ENCI 303 Lecture PS-19 Optimization 2
84 b Unidimensional Search Methods Most algorithms for unconstrained and constrained optimisation use an efficient unidimensional optimisation technique.
Optimization in Engineering Design Georgia Institute of Technology Systems Realization Laboratory 101 Quasi-Newton Methods.
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
Multivariate Unconstrained Optimisation First we consider algorithms for functions for which derivatives are not available. Could try to extend direct.
“On Sizing and Shifting The BFGS Update Within The Sized-Broyden Family of Secant Updates” Richard Tapia (Joint work with H. Yabe and H.J. Martinez) Rice.
CSE 245: Computer Aided Circuit Simulation and Verification Matrix Computations: Iterative Methods I Chung-Kuan Cheng.
559 Fish 559; Lecture 5 Non-linear Minimization. 559 Introduction Non-linear minimization (or optimization) is the numerical technique that is used by.
Quasi-Newton Methods of Optimization Lecture 2. General Algorithm n A Baseline Scenario Algorithm U (Model algorithm for n- dimensional unconstrained.
1cs426-winter-2008 Notes  Will add references to splines on web page.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Lecture 13. Geometry Optimization References Computational chemistry: Introduction to the theory and applications of molecular and quantum mechanics, E.
1 Chapter 6 General Strategy for Gradient methods (1) Calculate a search direction (2) Select a step length in that direction to reduce f(x) Steepest Descent.
Survey of unconstrained optimization gradient based algorithms
Hand-written character recognition
Data assimilation for weather forecasting G.W. Inverarity 06/05/15.
INTRO TO OPTIMIZATION MATH-415 Numerical Analysis 1.
Non-linear Minimization
Computational Optimization
Computational Optimization
CS5321 Numerical Optimization
Non-linear Least-Squares
Collaborative Filtering Matrix Factorization Approach
CS5321 Numerical Optimization
CS5321 Numerical Optimization
CS5321 Numerical Optimization
~ Least Squares example
~ Least Squares example
Performance Optimization
Section 3: Second Order Methods
CS5321 Numerical Optimization
Presentation transcript:

1cs542g-term Notes  Extra class this Friday 1-2pm  If you want to receive s about the course (and are auditing) send me

2cs542g-term From last time  Steepest Descent: Start with guess Until converged:  Find direction  Choose step size  Next guess is  Line search: keep picking different step sizes until satisfied (Reduce a multidimensional problem to 1D)

3cs542g-term Simplest line search  Start with initial step-size E.g. suggested by user, or from last step of algorithm  Check if it reduces f “enough”:  If not, halve the step and try again (Also, if first guess works for a few iterations in a row, try increasing step size)  Not enough to guarantee convergence, but often does OK in practice

4cs542g-term Convergence of Steepest Descent  We’ll use a model problem:  Here A is symmetric positive definite, so 0 is the unique minimum (f is strictly convex)  Gradient is:  We can further simplify: change to eigenvector basis, A becomes diagonal

5cs542g-term Convergence of SD cont’d  For the benefit of the doubt, assume line-search is “perfect”: picks the step to exactly minimize

6cs542g-term Diagnosis  Problem occurs for ill-conditioned A  Quite soon the bulk of the error is along the “shallow” directions, not the steepest (gradient)  Typical improved strategy: pick smarter directions Conjugate Gradient (later in the course): avoid repeating the same directions Newton and Quasi-Newton: try to pick the optimal direction

7cs542g-term Newton’s Method  Use the Hessian: second derivatives  Model the objective function as a quadratic  Minimize the model (solve a linear system)

8cs542g-term Newton’s Method  Now need to evaluate Hessian and gradient, and solve a linear system Perhaps too expensive…  But, can get quadratic convergence (# significant digits doubles each iteration)  But can also fail in more ways Hessian might be singular, indefinite, or otherwise unhelpful Higher-order nonlinearity might cause divergence Some of these problems can occur even if f is strictly convex

9cs542g-term Making Newton more robust  Modify Hessian to make it positive definite (e.g. add scaled identity): mixes in SD to guarantee descent direction (“regularization”)  Line search methods: use Newton direction, but add line search to make sure step is good  Trust region methods: only use quadratic model in a small “trust region”, don’t overstep bounds (and steadily shrink trust region to convergence)

10cs542g-term Quasi-Newton  Attack problems with Hessian (expense and possible ill-conditioning): build approximation to Hessian from information from gradients:  Example: BFGS (use this for low rank updates to approximation of H)