Presentation is loading. Please wait.

Presentation is loading. Please wait.

EE-M110 2006/7: EF L12&13 1/23, v2.0 Lectures 12&13: Persistent Excitation for Off-line and On-line Parameter Estimation Dr Martin Brown Room: E1k Email:

Similar presentations

Presentation on theme: "EE-M110 2006/7: EF L12&13 1/23, v2.0 Lectures 12&13: Persistent Excitation for Off-line and On-line Parameter Estimation Dr Martin Brown Room: E1k Email:"— Presentation transcript:

1 EE-M /7: EF L12&13 1/23, v2.0 Lectures 12&13: Persistent Excitation for Off-line and On-line Parameter Estimation Dr Martin Brown Room: E1k Telephone:

2 EE-M /7: EF L12&13 2/23, v2.0 Outline 13&14 Persistent excitation and identifiability 1.Structure of X T X 2.Role of signal magnitude 3.Role of signal correlation 4.Types of system identification signals for experimental design) On-line estimation and persistent excitation 5.On-line persistent excitation 6.Time-varying parameters 7.Exponential Recursive Least Squares (RLS)

3 EE-M /7: EF L12&13 3/23, v2.0 Resources 13&14 Core reading Ljung chapter 13 On-line notes, chapter 5 Norton, Chapter 8

4 EE-M /7: EF L12&13 4/23, v2.0 Central Question: Experimental Design An important part of system identification is experimental design Experimental design is involved with answering the question of how experiments should be constructed to to maximise the information collected with the minimum amount of effort/cost For system identification, this corresponds to how the input/control signal injected into the plant should be chosen to best identify the parameters N.B. This is relative to the model structure (i.e. different model structures will have different optimal model designs).

5 EE-M /7: EF L12&13 5/23, v2.0 What is Persistent Excitation Persistent excitation refers to the design of a signal, u(t), that produces estimation data D={X,y} which is rich enough to satisfactorily identify the parameters The parameter accuracy/covariance is determined by: Ideally, and E(x i 2 )>>  y 2 The variance/covariance can be made smaller (better) by: 1.Reducing the measurement error variance (hard) 2.Collecting more data (but this often costs money) 3.Make the signals larger (but there are physical limits) 4.Make the signals independent (difficult for dynamics)

6 EE-M /7: EF L12&13 6/23, v2.0 Review: X T X Matrix The variance/covariance matrix, X T X, (and its inverse) is central in many system identification/parameter estimation tasks Consider a model

7 EE-M /7: EF L12&13 7/23, v2.0 Identifying Parameters For a set of measured exemplars D={X,y}, there are several (related) concepts that determine how well the parameters can be estimated (off-line, in batch mode) 1, i.e. how well can the parameters be identified or equivalently, what is the region of uncertainty about the estimated values . 2Is (X T X) non-singular? i.e. can the normal equations be solved uniquely 3Are the parameter estimates significantly non-zero? All of these are related and influenced/determined by how the input data X is generated/collected. ^

8 EE-M /7: EF L12&13 8/23, v2.0 Example: Signal Magnitude & Noise Consider feeding steps of magnitude 0.01, 0.1 and 1 into the first order, electrical circuit with The magnitude of the signals strongly influences the identifiability of the parameters. Typically, each signal should be of similar magnitude and high in relation to the measurement noise.

9 EE-M /7: EF L12&13 9/23, v2.0 Example: Signals Interactions Consider collecting data from a model of the form: Each input is u i (t) = sin(0.5t), 20 samples: Note that X = [u 1 u 2 ] is singular Now consider u 1 (t) = sin(0.5t), u 2 (t) = cos(0.5t), E(u 1 u 2 )  0, E(u i 2 )=c The input signals are ~orthogonal This is difficult with feedback …

10 EE-M /7: EF L12&13 10/23, v2.0 Good and Bad Covariance Matrices Ideal structure of (X T X) -1 is which means that: Each parameter has the same variance, and the estimates are uncorrelated. In addition, if E(x i 2 )>>  y 2, the parameter variances are small. Each parameter can be identified to the same accuracy For modelling and control, we want to feed an input signal in produces a matrix with these properties.

11 EE-M /7: EF L12&13 11/23, v2.0 How to Measure Goodness? There are several ways to assess/compare how good a particular signal is: Cond(X T X) = max / min This measures the ratio of the maximum signal to the minimum signal correlations Smaller Cond(X T X) is better Cond(I) = 1 Choose u to min u Cond(X T X) Insensitive to the signal magnitude, just measures the degree of correlation Well determined = 12.8 = 0.52 poorly determined

12 EE-M /7: EF L12&13 12/23, v2.0 Signal Correlation and Dynamics So far, we have just discussed choosing input signals that are uncorrelated/orthogonal However, dynamics/feedback introduce correlation between individual signals (i.e. between u(t) and u(t-1) and y(t-1) and y(t)): E(y(t-1)u(t))  0 This is because y(t) is related to u(t), especially when they change slowly A stable plant will track (correlate with) the input signal Condition will be worse u(t)y(t)u(t)y(t)

13 EE-M /7: EF L12&13 13/23, v2.0 Example 1: Impulse/Step Signal Any linear system is completely identified by its impulse (or step) response – because convolution can be used to calculate the output. However, as shown in Slide 8, there are several aspects that may make this identification difficult 1.Magnitude of the step signal (relative to the noise) & impulse 2.Length of the transient period, relative to the steady state 3.Generation of the impulse/step signal which may be infeasible due to control magnitude and/or actuator dynamics limits 4.High correlation between u(t) and u(t-k), steady state adds little Note that if the plant model is non-linear, an impulse/step only collects information at one operating point, so if the aim is to reject non-linear components, step/impulse trains of different amplitudes must be used t u(t)u(t)u(t)u(t)u(t)u(t) tt

14 EE-M /7: EF L12&13 14/23, v2.0 Example 2: Sinusoidal Signal While a sinusoid may look to be a rich enough signal to identify linear models It can be used to identify the gain margin and phase advance for one particular frequency However, can only be used when the maximum control delay is 1, because u(t) =  1 u(t-1) +  2 u(t-2) Similar for the output feedback delay as well (because in the steady state, the output is also sinusoidal).

15 EE-M /7: EF L12&13 15/23, v2.0 Example 3: Random Signal A random signal is persistently exciting for a linear model of any order It involves a range of amplitudes and so can be used for non- linear terms as well. However, It is a bit of a “scatter gun” approach It can be wasteful when the model structure is reasonably well-known There may be limits on the actuator dynamics Difficult to use on-line, where the control action is “smooth”

16 EE-M /7: EF L12&13 16/23, v2.0 On-Line Parameter Estimation So far, it has been assumed that the parameter estimation is being performed off-line –Collect a fixed size data set –Estimate the parameters –Issues of parameter identifiability are related to a fixed data set On-line parameter estimation is more complex –Typically a plant is controlled to a set-point for a long period of time –The recursive calculation is often re-set after fixed intervals (re-set floating point errors) –Sometimes need to track time-varying parameters

17 EE-M /7: EF L12&13 17/23, v2.0 Time Varying Parameters One reason for considering on-line/recursive parameter estimation is to model systems where the linear parameters vary slowly with time Common parameter changes are step or slow drifts The aim is to treat the systems as slowly changing, and the model must be kept “plastic enough” to respond to changes in the parameters Note that, strictly speaking, this is now a non-linear system where the dynamics of the parameters are much slower than the dynamics of the system’s states.

18 EE-M /7: EF L12&13 18/23, v2.0 Long Term Convergence & Plasticity Using either the normal equations or the equivalent on-line, recursive version, when the amount of data increases, the parameter estimates tend to the true values and the effect of a new datum is close to zero. To model parametric drifts, the parameter estimates must include a term that makes the model more dependent on recent large residuals This can be achieved by defining a modified performance function where the residuals are weighted by a time decay factor

19 EE-M /7: EF L12&13 19/23, v2.0 Exponential RLS 1.Form the new input vector x(t+1) using the new data 2.Form  (t+1) from the model using 3.Form P(t+1) using 4.Update the least squares estimate Proceed with next time step

20 EE-M /7: EF L12&13 20/23, v2.0 Example: Exponential RLS Consider the first order electrical circuit example Here a and k are functions of time and both linearly vary between 1 and 2 during the length of the simulation Input signal is sinusoidal and noise N(0,0.01) is added There is a balance between noise filtering and model/parameter plasticity

21 EE-M /7: EF L12&13 21/23, v2.0 Parameter Convergence & Persistent Excitation While this algorithm is relatively simple, it has two important, related aspects that must be considered 1.What is the value of ? 2.What form of persistently exciting input is needed? When is 1, this is just standard RLS estimation. When <0.9, the model is extremely adaptive and the parameters will not generally converge when the measurement noise is significant As the model becomes more plastic, the input signal must be sufficiently persistently exciting over every significant time window to stop random parameter drift/premature convergence

22 EE-M /7: EF L12&13 22/23, v2.0 Summary 13&14 The engineer’s aim is to minimise the amount of data collected to identify the parameters sufficiently accurately Signal magnitude should be as large as possible to improve the signal/noise ratio and to minimize the parameter covariances. However, the signal should not to large enough to violate any system constraints or to make the unknown system significantly non-linear Signal type & frequency must be smooth enough not to exceed any dynamic constraints, however the dynamics must excite any potential dynamics. When parameter estimation is on-line, this imposes additional constraints as the signals must be sufficiently exciting for each time period Exponential-forgetting can be used to track time-varying parameters, but previous comments must hold

23 EE-M /7: EF L12&13 23/23, v2.0 Laboratory 13&14 1. Prove Slide 14 relationship for a sin function – what are  1 and  2 2. Measure the Cond(X T X) and the parameter estimates for: –Step –Sin –Random for the electrical simulation. Try varying the magnitudes of the step signal as well. 3. Implement the exponential RLS for the electrical simulation for time-varying parameters on Slide 20. Try changing the input/control signals and compare the responses.

Download ppt "EE-M110 2006/7: EF L12&13 1/23, v2.0 Lectures 12&13: Persistent Excitation for Off-line and On-line Parameter Estimation Dr Martin Brown Room: E1k Email:"

Similar presentations

Ads by Google