ECE 530 – Analysis Techniques for Large-Scale Electrical Systems

Slides:



Advertisements
Similar presentations
Linear Inverse Problems
Advertisements

Eigen Decomposition and Singular Value Decomposition
Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Autar Kaw Humberto Isaza
MATH 685/ CSI 700/ OR 682 Lecture Notes
Solution of linear system of equations
Chapter 9 Gauss Elimination The Islamic University of Gaza
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Chapter 5 Orthogonality
Curve-Fitting Regression
Math for CSLecture 41 Linear Least Squares Problem Over-determined systems Minimization problem: Least squares norm Normal Equations Singular Value Decomposition.
Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues
Linear Least Squares QR Factorization. Systems of linear equations Problem to solve: M x = b Given M x = b : Is there a solution? Is the solution unique?
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
1 Systems of Linear Equations Error Analysis and System Condition.
Matrices CS485/685 Computer Vision Dr. George Bebis.
Chapter 10 Real Inner Products and Least-Square (cont.)
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Introduction to MATLAB 7 for Engineers William J. Palm.
Boyce/DiPrima 9th ed, Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues Elementary Differential Equations and Boundary Value Problems,
Computational Methods in Physics PHYS 3437 Dr Rob Thacker Dept of Astronomy & Physics (MM-301C)
LU Decomposition 1. Introduction Another way of solving a system of equations is by using a factorization technique for matrices called LU decomposition.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
CHAPTER FIVE Orthogonality Why orthogonal? Least square problem Accuracy of Numerical computation.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
SVD: Singular Value Decomposition
Least SquaresELE Adaptive Signal Processing 1 Method of Least Squares.
Method of Least Squares. Least Squares Method of Least Squares:  Deterministic approach The inputs u(1), u(2),..., u(N) are applied to the system The.
Chapter 3 Solution of Algebraic Equations 1 ChE 401: Computational Techniques for Chemical Engineers Fall 2009/2010 DRAFT SLIDES.
Orthogonality and Least Squares
Chapter 7 Inner Product Spaces 大葉大學 資訊工程系 黃鈴玲 Linear Algebra.
Computing Eigen Information for Small Matrices The eigen equation can be rearranged as follows: Ax = x  Ax = I n x  Ax - I n x = 0  (A - I n )x = 0.
MATH 685/ CSI 700/ OR 682 Lecture Notes Lecture 4. Least squares.
Lesson 3 CSPP58001.
ECE 576 – Power System Dynamics and Stability Prof. Tom Overbye Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Multi-area Nonlinear State Estimation using Distributed Semidefinite Programming Hao Zhu October 15, 2012 Acknowledgements: Prof. G.
Chapter 9 Gauss Elimination The Islamic University of Gaza
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Lecture 6 - Single Variable Problems & Systems of Equations CVEN 302 June 14, 2002.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 576 – Power System Dynamics and Stability Prof. Tom Overbye Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Instructor: Mircea Nicolescu Lecture 8 CS 485 / 685 Computer Vision.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Singular Value Decomposition and Numerical Rank. The SVD was established for real square matrices in the 1870’s by Beltrami & Jordan for complex square.
Parameter estimation class 5 Multiple View Geometry CPSC 689 Slides modified from Marc Pollefeys’ Comp
Least Squares Problems From Wikipedia, the free encyclopedia The method of least squares is a standard approach to the approximate solution of overdetermined.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems
Chapter 12: Data Analysis by linear least squares Overview: Formulate problem as an over-determined linear system of equations Solve the linear system.
ECE 576 – Power System Dynamics and Stability Prof. Tom Overbye University of Illinois at Urbana-Champaign 1 Lecture 23: Small Signal.
Section 8 Numerical Analysis CSCI E-24 José Luis Ramírez Herrán October 20, 2015.
Lecture 9 Numerical Analysis. Solution of Linear System of Equations Chapter 3.
Let W be a subspace of R n, y any vector in R n, and the orthogonal projection of y onto W. …
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
ASEN 5070: Statistical Orbit Determination I Fall 2014
Chapter 10: Solving Linear Systems of Equations
Singular Value Decomposition SVD
ECEN 615 Methods of Electric Power Systems Analysis
ECEN 615 Methods of Electric Power Systems Analysis
Presentation transcript:

ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Lecture 20: Least-Squares Method Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign haozhu@illinois.edu 11/4/2014

Revisit LCDFs Closed line base case line k addition case

LCDF : Evaluation We can evaluate LCDF by reversing the line outage Recall how we define LODF outaged base case outage case

LCDF : Evaluation So the post-outage line flow 𝑓 𝑘 in LCDF is exactly the pre-outage line k flow in LODF But the change in line l flow in LCDF becomes the opposite of that in LODF So for the LCDF calculation Δ 𝑓 𝑙 =− 𝑑 𝑙 𝑘 𝑓 𝑘 Hence, we have 𝐿𝐶𝐷 𝐹 𝑙 𝑘 = Δ 𝑓 𝑙 𝑓 𝑘 =− 𝑑 𝑙 𝑘 We can verify this from the earlier example on 𝐿𝑂𝐷 𝐹 3 4 for the 5-bus case

Least Squares So far we have considered the solution of Ax = b in which A is a square matrix; as long as A is nonsingular there is a single solution That is, we have the same number of equations (m) as unknowns (n) Many problems are overdetermined in which there more equations than unknowns (m > n) Overdetermined systems are usually inconsistent, in which no value of x exactly solves all the equations Underdetermined systems have more unknowns than equations (m < n); they never have a unique solution but are usually consistent

Method of Least Squares The least squares method is a solution approach for determining an approximate solution for an overdetermined system If the system is inconsistent, then not all of the equations can be exactly satisfied The difference for each equation between its exact solution and the estimated solution is known as the error Least squares seeks to minimize the sum of the squares of the errors Weighted least squares allows differ weights for the equations

Least Squares Problem Consider 𝐀𝐱=𝐛, where 𝐀∈ 𝐑 𝑚×𝑛 , 𝐱∈ 𝐑 𝑛 , 𝐛∈ 𝐑 𝑚 or

Least Squares Solution We write (ai)T for the row i of A and ai is a column vector Here, m ≥ n and the solution we are seeking is that which minimizes 𝐀𝐱−𝐛 𝑝 , where p denotes some specific vector norm Since usually an overdetermined system has no exact solution, the best we can do is determine an x that minimizes the desired norm.

Example 1: Choice of p We discuss the choice of p in terms of a specific example Consider the equation Ax = b with Hence three equations and one unknown We will consider three possible choices for p

Example 1: Choice of p (i) p = 1 (ii) p = 2 (iii) p = 

The Least Squares Problem In general, is non-differentiable for p = 1 or p = ∞ The choice of p = 2 has become well established and given the least-squares fit interpretation We next motivate the choice of p = 2 by first considering the least–squares problem

The Least Squares Problem The problem is tractable for 2 major reasons (i) the function is differentiable in x ; and min 𝐱∈ 𝐑 𝑛 𝐀𝐱−𝐛 2

The Least Squares Problem (ii) the norm is preserved under orthogonal transformations: with Q an arbitrary orthogonal matrix; that is, Q satisfies 𝐐 𝐐 𝑇 = 𝐐 𝑇 𝐐=𝐈 𝐐∈ 𝐑 𝑛×𝑛

The Least Squares Problem We introduce next, the basic underlying assumption: A is of full (column) rank, i.e., the columns of A constitute a set of linearly independent vectors This assumption implies that the rank of A is n because n ≤ m since we are dealing with an overdetermined system Fact: The least squares problem solution x* satisfies

Proof of Fact Since by definition the least squares solution x* minimizes at the optimum, the derivative of this function vanishes:

Implications This underlying assumption implies that A is full column rank Therefore, the fact that ATA is positive definite (p.d.) follows from considering any x ≠ 0 and evaluating which is the definition of a p.d. matrix We use the shorthand ATA ≻ 0 for ATA being a symmetric, positive definite matrix

Implications The underlying assumption that A is full rank and therefore ATA is p.d. implies that there exists a unique least squares solution Note: we use the inverse in a conceptual, rather than a computational, sense The below formulation is known as the normal equations, with the conceptual solution as its unique solution

Implications An important implication of positive definiteness is that we can factor ATA since ATA ≻ 0 The expression ATA = GTG is called the Cholesky factorization of any symmetric p.d. matrix ATA

Least Squares Solution Algorithm Step 1: Compute the lower triangular part of ATA Step 2: Obtain the Cholesky Factorization Step 3: Compute Step 4: Solve for y using forward substitution in and for x using backward substitution in

Practical Considerations The two key problems that arise in practice with the triangularization procedure are: While A maybe sparse, ATA is much less sparse and consequently requires more computing and storage resources for the solution ATA may be numerically less well-conditioned than A 20

Example 2: Loss of Sparsity Assume the B matrix for a network is Then BTB is Second neighbors are now connected! But large networks are still sparse, just not as sparse

Numerical Conditioning To understand the numerical ill-conditioning issue, we need to introduce terminology We define the norm of a matrix 𝐁∈ 𝐑 𝑚×𝑛 to be

Numerical Conditioning 𝐁 i.e., li is a root of the polynomial In words, the norm of matrix B is the square root of the largest eigenvalue of BTB

Numerical Conditioning The conditioning number of a matrix B is defined as A well–conditioned matrix has a small value of , close to 1 The larger the value of is, the more pronounced the ill-conditioning is The ill-conditioned nature of ATA may severely impact the accuracy of the computed solution

Example 3: Ill-Conditioned ATA We illustrate the fact that an ill-conditioned matrix ATA results in highly sensitive solutions of least-squares problems in this example with then

Example 3: Ill-Conditioned ATA We introduce a “noise” in A to be the matrix dA

Example 3: Ill-Conditioned ATA This “noise” leads to the error E in the computation of ATA with Let and assume that there is no “noise” in , i.e.,

Example 3: Ill-Conditioned ATA The resulting error in solving the normal equations is independent of since it is caused purely by 𝜹𝑨 and 𝜹𝑨 𝑇 𝒃 Let x be the true solution of the normal equations ⇒ 𝒙= 1 0

Example 3: Ill-Conditioned ATA Let be the solution of the system with the error arising due to , i.e., the solution of Therefore,

Example 3: Ill-Conditioned ATA Therefore, the relative error is Now, the conditioning number and So the product approximates the relative error

Example 3: Ill-Conditioned ATA Thus the conditioning number is a major contributor to the error in the least-squares solution In other words, the sensitivity of the solution to any system error, be its data entry or of a numerical nature, is very dependent on

Solving the Least-Squares Problem With the previous background we proceed to the typical schemes in use for solving least squares problems, all along paying adequate attention to the numerical aspects of the solution approach If the matrix is full, then often the best solution approach is to use a singular value decomposition (SVD), to form a matrix known as the pseudo-inverse of the matrix We'll cover this later after first considering the sparse problem We first review some fundamental building blocks and then present the key results useful for the sparse matrices common in state estimation

Power System State Estimation Power system state estimation (SE) is covered in ECE 573, so we'll just touch on it here; it is a key least squares application Overall goal is to come up with a power flow model for the present "state" of the power system based on the actual system measurements SE assumes the topology and parameters of the transmission network are mostly known Measurements come from SCADA, and increasingly, from PMUs

Power System State Estimation Good introductory reference is Power Generation, Operation and Control by Wood, Wollenberg and Sheble, 3rd Edition Problem can be formulated in a nonlinear, weighted least squares form as where J(x) is the cost function, x are the state variables (primarily bus voltage magnitudes and angles), {zi} are the m measurements, f(x) relates the states to the measurements and i is the assumed standard deviation

Measurement Example Assume we measure the real and reactive power flowing into one end of a transmission line; then the zi-fi(x) functions correspond to Two measurements for four unknowns Other measurements, such as the flow at the other end, power injection and voltage magnitudes, add redundancy

Assumed Error Hence the goal is to decrease the error between the measurements and the assumed model states x The i term weighs the various measurements, recognizing that they can have vastly different assumed errors Measurement error is assumed Gaussian; whether it is or not is another question; outliers (bad measurements) are often removed