SYSTEMS Identification

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements


ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The Linear Prediction Model The Autocorrelation Method Levinson and Durbin.
Pattern Recognition and Machine Learning
CHAPTER 3 CHAPTER 3 R ECURSIVE E STIMATION FOR L INEAR M ODELS Organization of chapter in ISSO –Linear models Relationship between least-squares and mean-square.
Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
Extended Kalman Filter (EKF) And some other useful Kalman stuff!
CY3A2 System identification Modelling Elvis Impersonators Fresh evidence that pop stars are more popular dead than alive. The University of Missouri’s.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: The FIR Adaptive Filter The LMS Adaptive Filter Stability and Convergence.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Newton’s Method Application to LMS Recursive Least Squares Exponentially-Weighted.
Lecture 11: Recursive Parameter Estimation
The Simple Linear Regression Model: Specification and Estimation
280 SYSTEM IDENTIFICATION The System Identification Problem is to estimate a model of a system based on input-output data. Basic Configuration continuous.
Maximum likelihood (ML) and likelihood ratio (LR) test
Factor Analysis Purpose of Factor Analysis
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Curve-Fitting Regression
SYSTEMS Identification
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
SYSTEMS Identification
Development of Empirical Models From Process Data
Prediction and model selection
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Estimation and the Kalman Filter David Johnson. The Mean of a Discrete Distribution “I have more legs than average”
Course AE4-T40 Lecture 5: Control Apllication
Linear and generalised linear models
Linear and generalised linear models
Kalman Filtering Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read.
SYSTEMS Identification
Maximum likelihood (ML)
Multivariable Control Systems Ali Karimpour Assistant Professor Ferdowsi University of Mashhad.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Adaptive Signal Processing
Normalised Least Mean-Square Adaptive Filtering
RLSELE Adaptive Signal Processing 1 Recursive Least-Squares (RLS) Adaptive Filters.
Algorithm Taxonomy Thus far we have focused on:
Particle Filtering in Network Tomography
CORRELATION & REGRESSION
4. Linear optimal Filters and Predictors 윤영규 ADSLAB.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 Chapter 2 1. Parametric Models. 2 Parametric Models The first step in the design of online parameter identification (PI) algorithms is to lump the unknown.
CHAPTER 4 Adaptive Tapped-delay-line Filters Using the Least Squares Adaptive Filtering.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Generalised method of moments approach to testing the CAPM Nimesh Mistry Filipp Levin.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
Non-Linear Models. Non-Linear Growth models many models cannot be transformed into a linear model The Mechanistic Growth Model Equation: or (ignoring.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Normal Equations The Orthogonality Principle Solution of the Normal Equations.
Introducing Error Co-variances in the ARM Variational Analysis Minghua Zhang (Stony Brook University/SUNY) and Shaocheng Xie (Lawrence Livermore National.
Feedback Controllers Chapter 8
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad.
State-Space Recursive Least Squares with Adaptive Memory College of Electrical & Mechanical Engineering National University of Sciences & Technology (NUST)
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Thomas F. Edgar (UT-Austin) RLS – Linear Models Virtual Control Book 12/06 Recursive Least Squares Parameter Estimation for Linear Steady State and Dynamic.
STATISTICAL ORBIT DETERMINATION Kalman (sequential) filter
CHAPTER 3 RECURSIVE ESTIMATION FOR LINEAR MODELS
Instructor :Dr. Aamer Iqbal Bhatti
SYSTEMS Identification
Feedback Controllers Chapter 8
16. Mean Square Estimation
Presentation transcript:

SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad <<<1.1>>> ###Control System Design### {{{Control, Design}}} Reference: “System Identification Theory For The User” Lennart Ljung

Recursive estimation methods Lecture 11 Recursive estimation methods Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

Introduction Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

Introduction In many cases it is necessary, or useful, to have a model of the system available on-line. The need for such an on-line model is required in order to: Which input should be applied at the next sampling instant? Adaptive How should the parameters of a matched filter be tuned? What are the best predictions of the next few output? Has a failure occurred and, if so, of what type? Adaptive control, adaptive filtering, adaptive signal processing, adaptive prediction.

Introduction The on-line computation of the model must completed during one sampling interval. Identification techniques that comply with this requirement will be called: Recursive identification methods. Used in this Reference. Recursive identification methods. On-line identification. Real-time identification. Adaptive parameter estimation. Sequential parameter estimation.

Introduction

Minimizing argument of some function or… Introduction Algorithm format Minimizing argument of some function or… General identification method: This form cannot be used in a recursive algorithm, since it cannot be completed in one sampling instant. Instead following recursive algorithm must comply: Information state Since the information in the latest pair of measurement { y(t) , u(t) } normally is small compared to the pervious information so there is a more suitable form Small numbers reflecting the relative information value in the latest measurement.

The Recursive Least-Squares Algorithm Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

The Recursive Least-Squares Algorithm Weighted LS Criterion The estimate for the weighted least squares is: Where

The Recursive Least-Squares Algorithm Recursive algorithm Suppose the weighting sequence has the following property: Now

The Recursive Least-Squares Algorithm Recursive algorithm Suppose the weighting sequence has the following property:

The Recursive Least-Squares Algorithm Recursive algorithm Version with Efficient Matrix Inversion To avoid inverting at each step, let introduce Remember matrix inversion lemma

The Recursive Least-Squares Algorithm Version with Efficient Matrix Inversion Moreover we have We can summarize this version of algorithm as:

The Recursive Least-Squares Algorithm The size of the matrix R(t) will depend on the λ(t) Normalized Gain Version

The Recursive Least-Squares Algorithm Initial Condition A possibility could be to initialize only at a time instant t0 By LS method Clearly if P0 is large or t is large, then above estimate is the same as:

The Recursive Least-Squares Algorithm Asymptotic Properties of the Estimate

The Recursive Least-Squares Algorithm Multivariable case Remember SISO Now for MIMO

The Recursive Least-Squares Algorithm Kalman Filter Interpretation The Kalman Filter for estimating the state of system Kalman Filter The linear regression model can be cast to above form as: Now, let Exercise: Derive the Kalman filter for above mention system, and show that it is exactly same as the Recursive Least-Squares Algorithm for multivariable case.

The Recursive Least-Squares Algorithm Kalman Filter Interpretation Kalman filter interpretation gives important information, as well as some practical hints:

The Recursive Least-Squares Algorithm Coping with Time-varying Systems An important reason for using adaptive methods and recursive identification in practice is: The properties of the system may be time varying. We want the identification algorithm to track the variation. This is handled by weighted criterion, by assigning less weight to older measurements

The Recursive Least-Squares Algorithm Coping with Time-varying Systems These choices have the natural effect that in the recursive algorithms the step size will not decrease to zero.

The Recursive Least-Squares Algorithm Coping with Time-varying Systems Another and more formal alternative to deal with time-varying parameters is that the true parameters varies like a random walk so Exercise: Derive the Kalman filter for above mention system, and show that it is exactly same as the Recursive Least-Squares Algorithm for multivariable case. Note: The additive term R1(t) in P(t) prevents the gain L(t) from tending to zero.

The Recursive IV Method Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

The Recursive IV Method Remember Weighted LS Criterion: Where The IV estimate for instrumental variable method is:

The Recursive IV Method The IV estimate for instrumental variable method is:

Recursive Prediction-Error Methods Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

Recursive Prediction-Error Methods Analogous to the weighted LS case, let us consider a weighted quadratic prediction-error criterion Where so we have the gradient with respect to θ is

Recursive Prediction-Error Methods Analogous to the weighted LS case, let us consider a weighted quadratic prediction-error criterion Remember the general search algorithm developed for PEM as: For each iteration i, we collect one more data point, so now define As an approximation let:

Recursive Prediction-Error Methods Analogous to the weighted LS case, let us consider a weighted quadratic prediction-error criterion As an approximation let: With above approximation and taking μ(t)=1, we thus arrive at the algorithm: This terms must be recursive too.

Recursive Prediction-Error Methods This terms must be recursive too.

Recursive Prediction-Error Methods This terms must be recursive too.

Recursive Prediction-Error Methods This terms must be recursive too.

Recursive Prediction-Error Methods Family of recursive prediction error methods According to the model structure Wide family of methods According to the choice of R We shall call “RPEM” For example, the linear regression This is recursive least square method If we consider R(t)=I Where the gain could be normalized so This scheme has been widely used, under the name least mean squares (LMS)

Recursive Prediction-Error Methods Example 11.1 Recursive Maximum Likelihood Consider ARMAX model where and Remember chapter 10 By rule 11.41 This scheme is known as recursive maximum likelihood (RML)

Recursive Prediction-Error Methods Projection into DM The model structure is well defined only for giving stable predictors. In off-line minimization this must be kept in mind as a constraint. The same is true for the recursive minimization.

Recursive Prediction-Error Methods Asymptotic Properties The recursive prediction-error method is designed to make updates of θ in a direction that “on the average” is modified negative gradient of i.e.

Recursive Prediction-Error Methods Asymptotic Properties Moreover (see appendix 11a), for Gauss-Newton RPEM, with It can be shown that has an asymptotic normal distribution, which coincides with that of the corresponding off-line estimate. We thus have

Recursive Pseudolinear Regressions Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

Recursive Pseudolinear Regressions Consider the pseudo linear representation of the prediction And recall that this model structure contains, among other models, the general linear SISO model: A bootstrap method for estimating θ was given by (Chapter 10, 10.64) By Newton - Raphson method

Recursive Pseudolinear Regressions By Newton - Raphson method

Recursive Pseudolinear Regressions

Recursive Pseudolinear Regressions Family of RPLRs The RPLR scheme represents a family of well-known algorithms when applied to different special cases of The RPLR scheme represents a family of well-known algorithms when applied to different special cases of The ARMAX case is perhaps the best known of this. If we choose This scheme is known as extended least squares (ELS). Other special cases are displayed in following table:

Recursive Pseudolinear Regressions Other special cases are displayed in following table:

The Choice of Updating Step Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

The Choice of Updating Step Recursive Prediction-Error Methods is based prediction error approach: Recursive Pseudolinear Regressions is based on correlation approach:

The Choice of Updating Step Recursive Prediction-Error Methods (RPEM) Recursive Pseudolinear Regressions (RPLR) The difference between prediction error approach and correlation approach is: Now we are going to speak about that modifies the update direction and determines the length of the update step. We just speak about RPEM, RPLR is the same just one must change

The Choice of Updating Step Update direction There are two basic choices of update directions: Better convergence rate Easier computation

The Choice of Updating Step Update Step: Adaptation gain An important aspect of recursive algorithm is, their ability to cope with time varing systems. There are two different ways of achieving this

The Choice of Updating Step Update Step: Adaptation gain In either case, the choice of update step is a trade-off between Tracking ability Noise sensitivity A high gain means that the algorithm is alert in tracking parameter changes but at the same time sensitive to disturbance in data.

The Choice of Updating Step Choice of forgetting factor The choice of forgetting profile β(t,k) is conceptually simple. For a system that changes gradually and in a stationary manner the most common choice is: The constant λ is always chosen slightly less than 1 so This means that measurements that are older than T0 samples are included in the criterion with a weight e-1≈0.36% of the most recent measurement. So T0 is the memory time constant. So we could select λ such that 1/(1-λ) reflects the ratio between the time constant of variations in the dynamics and those of the dynamics itself. Typical choices of λ are in the range between 0.98 and 0.995. For a system that undergoes sudden changes, rather than steady and slow ones, it is suitable to decrease λ(t) to a small value and then increase it to a value close to 1 again.

The Choice of Updating Step Choice of Gain γ(t)

The Choice of Updating Step Including a model of parameter changes Kalman Filter Interpretation Remember Now let

The Recursive Least-Squares Algorithm Kalman Filter Interpretation In the case of a linear regression model, this algorithm does give the optimal trade-off between tracking ability and noise sensitivity, in terms of parameter error covariance matrix. The case where the parameters are subject to variations that themselves are of a nonstationary nature, [i.e. R1(t) varies with t] needs a parallel algorithm. (see Anderson 1985)

The Choice of Updating Step Constant systems

The Choice of Updating Step Asymptotic behavior in the time-varying case

Implementation Topics to be covered include: Introduction. The Recursive Least-Squares Algorithm. The Recursive IV Method. Recursive Prediction-Error Methods. Recursive Pseudolinear Regressions. The Choice of Updating Step. Implementation.

Implementation Implementation The basic, general Gauss-Newton algorithm was given in RPEM and RPLR. Recursive Prediction-Error Methods (RPEM) Recursive Pseudolinear Regressions (RPLR) Inverse manipulation is not suited for direct implementation, (d*d matrix R must be inverted) We shall discuss some aspects on how to best implement recursive algorithm. By using matrix inversion lemma Here η (a d*p matrix) represents either φ or ψ depending on the approcach. But this is p*p

Implementation Implementation Unfortunately, the P-recursion which in fact is a Riccati equation is not numerically sound: the equation is sensitive to round-off errors that can accumulate and make P(t) indefinite Using factorization It is useful to represent the data matrices in factorized form so as to work with better conditioned matrices. Cholesky decomposition, which Q(t) is triangular UD-decomposition, which U(t) is triangular and D is diagonal Here we shall give some details of a related algorithm, which is directed based on Householder transformation (problem 10T.1). (by Morf and Kailath)

Implementation Using factorization Step 1: Let Form (p+d)*(p+d) matrix Step 2: Apply an orthogonal (p+d)*(p+d) transformation T (TTT=I) to L(t-1) so that TL(t-1) becomes an upper triangular matrix. (Use QR-factorization) Let to partition TL(t-1) as:

Implementation Using factorization

Implementation Using factorization Step 3: Now L(t) and P(t) are:

Implementation Using factorization in summary Now L(t) and P(t) are: There are several advantages with this particular way of performing. The only essential computation to perform is the triangularization step. This step gives new Q and the gain L after simple additional calculations. Note that Π(t) is triangular p*p matrix, so it is easy to invert it. Condition number of the matrix L(t-1) is much better than that of P.