ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.

Slides:



Advertisements
Similar presentations
Optimization.
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Optimization of thermal processes
Optimization 吳育德.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
Steepest Decent and Conjugate Gradients (CG). Solving of the linear equation system.
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
Jonathan Richard Shewchuk Reading Group Presention By David Cline
Gradient Methods April Preview Background Steepest Descent Conjugate Gradient.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Function Optimization Newton’s Method. Conjugate Gradients
Tutorial 12 Unconstrained optimization Conjugate gradients.
Gradient Methods May Preview Background Steepest Descent Conjugate Gradient.
Tutorial 5-6 Function Optimization. Line Search. Taylor Series for Rn
Gradient Methods Yaron Lipman May Preview Background Steepest Descent Conjugate Gradient.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
Function Optimization. Newton’s Method Conjugate Gradients Method
Monica Garika Chandana Guduru. METHODS TO SOLVE LINEAR SYSTEMS Direct methods Gaussian elimination method LU method for factorization Simplex method of.
Orthogonality and Least Squares
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
CSE 245: Computer Aided Circuit Simulation and Verification

9 1 Performance Optimization. 9 2 Basic Optimization Algorithm p k - Search Direction  k - Learning Rate or.
Numerical Analysis 1 EE, NCKU Tien-Hao Chang (Darby Chang)
Computational Optimization
UNCONSTRAINED MULTIVARIABLE
Systems of Linear Equations Iterative Methods
Eigenvalue Problems Solving linear systems Ax = b is one part of numerical linear algebra, and involves manipulating the rows of a matrix. The second main.
Chapter 5: The Orthogonality and Least Squares
ENCI 303 Lecture PS-19 Optimization 2
1 Unconstrained Optimization Objective: Find minimum of F(X) where X is a vector of design variables We may know lower and upper bounds for optimum No.
1 Optimization Multi-Dimensional Unconstrained Optimization Part II: Gradient Methods.
Chapter 5 Eigenvalues and Eigenvectors 大葉大學 資訊工程系 黃鈴玲 Linear Algebra.
CSE 245: Computer Aided Circuit Simulation and Verification Matrix Computations: Iterative Methods I Chung-Kuan Cheng.
Chapter 10 Real Inner Products and Least-Square
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
ECE 576 – Power System Dynamics and Stability Prof. Tom Overbye Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Case Study in Computational Science & Engineering - Lecture 5 1 Iterative Solution of Linear Systems Jacobi Method while not converged do { }
Elementary Linear Algebra Anton & Rorres, 9 th Edition Lecture Set – 07 Chapter 7: Eigenvalues, Eigenvectors.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Chapter 10 Minimization or Maximization of Functions.
ECE 476 Power System Analysis Lecture 12: Power Flow Prof. Tom Overbye Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Krylov-Subspace Methods - I Lecture 6 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
4.8 Rank Rank enables one to relate matrices to vectors, and vice versa. Definition Let A be an m  n matrix. The rows of A may be viewed as row vectors.
Solving Scalar Linear Systems A Little Theory For Jacobi Iteration
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Network Systems Lab. Korea Advanced Institute of Science and Technology No.1 Maximum Norms & Nonnegative Matrices  Weighted maximum norm e.g.) x1x1 x2x2.
MA5233 Lecture 6 Krylov Subspaces and Conjugate Gradients Wayne M. Lawton Department of Mathematics National University of Singapore 2 Science Drive 2.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Krylov-Subspace Methods - II Lecture 7 Alessandra Nardi Thanks to Prof. Jacob White, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Matrices CHAPTER 8.9 ~ Ch _2 Contents  8.9 Power of Matrices 8.9 Power of Matrices  8.10 Orthogonal Matrices 8.10 Orthogonal Matrices 
Conjugate gradient iteration One matrix-vector multiplication per iteration Two vector dot products per iteration Four n-vectors of working storage x 0.
The Landscape of Sparse Ax=b Solvers Direct A = LU Iterative y’ = Ay Non- symmetric Symmetric positive definite More RobustLess Storage More Robust More.
Krylov-Subspace Methods - I
Solving Poisson Equations Using Least Square Technique in Image Editing Colin Zheng Yi Li.
CSE 245: Computer Aided Circuit Simulation and Verification
CSE 245: Computer Aided Circuit Simulation and Verification
CSE 245: Computer Aided Circuit Simulation and Verification
Conjugate Gradient Method
CS5321 Numerical Optimization
Administrivia: November 9, 2009
Performance Optimization
Conjugate Direction Methods
Presentation transcript:

ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign 12/2/ Lecture 25: Krylov Subspace Methods

Announcements No class on Thursday Dec 4 Homework 8 posted, due on Thursday Dec 11 2

Krylov Subspace Outline Review of fields and vector spaces Eigensystem basics Definition of Krylov subspaces and annihilating polynomial Generic Krylov subspace solver Steepest descent Conjugate gradient 3

Krylov Subspace Iterative methods to solve Ax=b build on the idea that Given a matrix A and a vector v, the i th order Krylov subspace is defined as For a specified matrix A and a vector v, the largest value of i is bounded 4

Generic Krylov Subspace Solver The following is a generic Krylov subspace solver method for solving Ax = b using only matrix vector multiplies Step 1: Start with an initial guess x (0) and some predefined error tolerance  > 0; compute the residual, r (0) = b – A x (0) ; set i = 0 Step 2: While ||r (i) ||   Do (a) i := i + 1 (b) get K i (r (0),A) (c) find x (i) in {x (0) + K i (r (0),A)} to minimize ||r (i) || Stop 5

Note that no calculations are performed in Step 2 once i becomes greater than its largest value The Krylov subspace methods differ from each other in – the construction scheme of the Krylov subspace in Step 2(b) of the scheme – the residual minimization criterion used in Step 2(c) A common initial guess is x (0) = 0, giving r (0) = b – A x (0) = b Every solver involves the A matrix only in matrix- vector products: A i r (0), i=1,2,… Krylov Subspace Solver 6

Directly constructing the Krylov Subspace for any A and r (0) would be computationally expensive We will instead introduce iterative optimization methods for solving Ax = b, which turns out to be a a special case of Krylov Subspace method Without loss of generality, consider the system Ax = b where A is symmetric (i.e., A = A T ) and positive definite (i.e., A ≻0, all eigenvalues nonnegative) Any Ax = b with nonsingular A is equivalent to A T Ax = A T b where A T A is symmetric and positive definite Iterative Optimization Methods 7

Optimization Problem 8

Steepest Descent Algorithm 9 Note there is only one matrix, vector multiply per iteration

Steepest Descent Convergence 10

Steepest Descent Convergence 11

An improvement over the steepest descent is to take the exact number of steps using a set of search directions and obtain the solution after n such steps; this is the basic idea in the conjugate direction methods Image compares steepest descent with a conjugate direction approach Conjugate Direction Methods Image Source: 12

The basic idea is the n search directions denoted by need to be A-orthogonal, that is At the i th iteration, we will update Conjugate Direction Methods 13

Stepsize Selection 14

Convergence Proof 15

Linearly Independent Directions Proposition: If A is positive definite, and the set of nonzero vectors d (0), d (1) … d (n-1) are, then these vectors are linearly independent (l.i.) Proof: Suppose there are constants  i, i=0,1,2,…n such Multiplying by A and then scalar product with d (i) gives Since A is positive definite, it follows  i = 0 Hence, the vectors are l.i. 16 Recall l.i. only if all  's = 0

Conjugate Direction Method 17 What we have not yet covered is how to get the n search directions. We'll cover that shortly, but the next slide presents an algorithm, followed by an example.

Orthogonalization To quickly generate A –orthogonal search directions, one can use the Gram-Schmidt orthogonalization procedure Suppose we are given a l.i. set of n v ectors {u 0, u 1, …, u n-1 }, successively construct d (j), j=0, 1, 2, … n-1, by removing from u j all the components along directions The trick is to use the gradient directions; i.e., u i = r (i) for all i=0,1,…,n-1, which yields the very popular conjugate gradient method 18

Conjugate Gradient Method 19

Conjugate Gradient Algorithm 20 Note that there is only one matrix vector multiply per iteration!

Conjugate Gradient Example Using the same system as before, let Select i=0, x (0) = 0,  = 0.1, then r (0) = b With i = 0, d (0) = r (0) = b 21

Conjugate Gradient Example 22 This first step exactly matches Steepest Descent

Conjugate Gradient Example With i=1 solve for  (1) Then

Conjugate Gradient Example And

Conjugate Gradient Example With i=2 solve for  (2) Then

Conjugate Gradient Example And 26 Done in 3 = n iterations!

Krylov Subspace Method 27

References D. P. Bertsekas, Nonlinear Programming, Chapter 1, 2 nd Edition, Athena Scientific, 1999 Y. Saad, Iterative Methods for Spare Linear Systems, 2002, free online at 28