6.5 Taylor Series Linearization

Slides:



Advertisements
Similar presentations
Numerical Solution of Linear Equations
Advertisements

Lecture 5 Newton-Raphson Method
CSE 330: Numerical Methods
Linear Algebra Applications in Matlab ME 303. Special Characters and Matlab Functions.
Linear Systems of Equations
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Computer Vision Optical Flow
Jonathan Richard Shewchuk Reading Group Presention By David Cline
Chapter 4 Roots of Equations
Function Optimization Newton’s Method. Conjugate Gradients
Curve-Fitting Regression
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Martin Mendez UASLP Chapter 61 Unit II.
Lecture Notes Dr. Rakhmad Arief Siregar Universiti Malaysia Perlis
Why Function Optimization ?
1 Systems of Linear Equations Error Analysis and System Condition.
Principles of Least Squares
Table of Contents Solving Systems of Linear Equations - Gaussian Elimination The method of solving a linear system of equations by Gaussian Elimination.
Least-Squares Regression
Taylor Series.
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Modern Navigation Thomas Herring
1 Ch. 4 Linear Models & Matrix Algebra Matrix algebra can be used: a. To express the system of equations in a compact manner. b. To find out whether solution.
Curve-Fitting Regression
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Elliptic PDEs and the Finite Difference Method
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Boundary Value Problems l Up to this point we have solved differential equations that have all of their initial conditions specified. l There is another.
Linear Systems – Iterative methods
Circuits Theory Examples Newton-Raphson Method. Formula for one-dimensional case: Series of successive solutions: If the iteration process is converged,
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Solving Non-Linear Equations (Root Finding)
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
Lecture 6 - Single Variable Problems & Systems of Equations CVEN 302 June 14, 2002.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Part 3 Chapter 12 Iterative Methods
MAT 2401 Linear Algebra 2.5 Applications of Matrix Operations
Copyright ©2015 Pearson Education, Inc. All rights reserved.
Computers in Civil Engineering 53:081 Spring 2003 Lecture #8 Roots of Equations: Systems of Equations.
Linear Systems Numerical Methods. 2 Jacobi Iterative Method Choose an initial guess (i.e. all zeros) and Iterate until the equality is satisfied. No guarantee.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
Chapter 7: Systems of Equations and Inequalities; Matrices
STATISTICAL ORBIT DETERMINATION Kalman (sequential) filter
Gauss-Siedel Method.
Numerical Analysis Lecture12.
Advanced Numerical Methods (S. A. Sahu) Code: AMC 51151
Lecture Slides Elementary Statistics Thirteenth Edition
Autar Kaw Benjamin Rigsby
CSE Differentiation Roger Crawfis.
Modelling data and curve fitting
The regression model in matrix form
Numerical Analysis Lecture 16.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Lial/Hungerford/Holcomb/Mullins: Mathematics with Applications 11e Finite Mathematics with Applications 11e Copyright ©2015 Pearson Education, Inc. All.
Instructor :Dr. Aamer Iqbal Bhatti
Nonlinear regression.
Numerical Analysis Lecture14.
6.2 Grid Search of Chi-Square Space
Numerical Analysis Lecture13.
5.2 Least-Squares Fit to a Straight Line
5.4 General Linear Least-Squares
6.1 Introduction to Chi-Square Space
6.6 The Marquardt Algorithm
topic4: Implicit method, Stability, ADI method
6.3 Gradient Search of Chi-Square Space
topic4: Implicit method, Stability, ADI method
Section 3: Second Order Methods
Numerical Analysis Lecture11.
Pivoting, Perturbation Analysis, Scaling and Equilibration
Presentation transcript:

6.5 Taylor Series Linearization writing out the Taylor series expansion simplification using the b and a notation expansion for the Gaussian peak example iterative solution of the matrix least-squares errors in the coefficients are given by the a-1 variance/covariance matrix tolerance toward initial guesses improving performance with a second-order expansion 6.5 : 1/8

The Approach Linearize the function locally by Taylor series expansion about a set of guess values for the coefficients. Use the linear least squares matrix method to calculate b and a and solve for a. Use the values of a as the new guess values for another Taylor series expansion Iterate to convergence. 6.5 : 1/8

Linearization of Coefficients Consider a function non-linear in its coefficients, y = f(x,a), where a is a vector of coefficients: a0, a1, ... , an. The function can be linearized by a first-order Taylor series expansion about the true coefficient values. Let the true values be denoted as m0, m1, ... , mn, and be represented by the vector, m. Simplify this expression by using the following variable changes. The linear approximation can now be written in a compact form. 6.5 : 2/8

We have seen this problem before! 6.5 : 1/8

Iterative Matrix Solution By defining, it is possible to recover a matrix formalism, where a-1 is the variance/covariance matrix. If the parameter values at the minimum in chi-square were substituted for m, the vector d would contain all zeros. In practice the parameters are estimated and used to construct an initial guess at the m vector. The expansion is performed around the guesses. The least-squares solution then yields a non-zero value for d. These two vectors are then used to construct a refined set of guesses, m'. The new vector, m', is used as a new set of guesses and the matrix calculation is repeated. The process stops when all dr ~ 0. 6.5 : 3/8

Example Gaussian Linearization Consider the Gaussian example we have been using, where a is a vector containing the coefficients. The function y(x,a) is used to compute Dy given on an earlier slide. The calculation of the a matrix and b vector require the following partial derivatives. 6.5 : 4/8

Iterative Solution The initial guesses used with the grid and gradient search algorithms were used. They appear in the table below as iteration 1. The computed parameters were then used as guesses for the second computation. They are iteration 2. Etc. a Iteration 1 2 3 4 g0 d0 2.000 +0.137 2.137 +0.004 2.141 0.000 g1 d1 51.000 -0.058 50.942 -0.005 50.937 +0.001 50.938 The final coefficients are essentially the same as those obtained with the grid and gradient search algorithms. 6.5 : 5/8

Coefficient Errors The error in the fit is given by the following with a0 = 2.141 and a1 = 50.938. There were eleven data values. The a-1 variance/covariance equation has the following two diagonal elements. The standard deviation of the two coefficients can now be calculated. These values are comparable to those obtained by fitting chi-square space to a two-dimensional parabola: sa0 = 0.034, sa1 = 0.045. 6.5 : 6/8

Tolerance Toward Initial Guesses The Taylor series linearization does not tolerate arbitrarily bad initial guesses. Consider the example Gaussian. If the value of a1 is fixed at 50.938, the smallest value of a0 that converged to the minimum in chi-square space was 0.25. The largest value of a0 that converged was 5.1. Values smaller than 0.25 and larger than 5.1 gave iterative values that diverged. These values are shown below. In contrast, the gradient search converged with values as small as 0.1. Values as large as 1,000 converged, albeit, taking ~1 min. 6.5 : 7/8

Tolerance Toward Initial Guesses The curvature of c2-space dictates whether the Taylor series approach will converge or diverge. You may recall, the elements of the matrix a are essentially second derivatives of c2-space. When those invert in sign, the Taylor series approach leads away from the minimum rather than toward it. 6.5 : 7/8