12.540 Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room 54-820A; 253-5941

Slides:



Advertisements
Similar presentations
Principles of the Global Positioning System Lecture 12 Prof. Thomas Herring Room A;
Advertisements

Lecture 13 L1 , L∞ Norm Problems and Linear Programming
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Integration of sensory modalities
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Variance and covariance M contains the mean Sums of squares General additive models.
The Simple Linear Regression Model: Specification and Estimation
AGC DSP AGC DSP Professor A G Constantinides© Estimation Theory We seek to determine from a set of data, a set of parameters such that their values would.
Chapter 4 Multiple Regression.
Adjustment theory / least squares adjustment Tutorial at IWAA2010 / examples Markus Schlösser adjustment theory Hamburg,
Course AE4-T40 Lecture 5: Control Apllication
Linear and generalised linear models
Basics of regression analysis
Lecture 3: Inferences using Least-Squares. Abstraction Vector of N random variables, x with joint probability density p(x) expectation x and covariance.
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Modern Navigation Thomas Herring
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Adaptive Signal Processing
Principles of the Global Positioning System Lecture 16 Prof. Thomas Herring Room A;
Principles of the Global Positioning System Lecture 13 Prof. Thomas Herring Room A;
Colorado Center for Astrodynamics Research The University of Colorado STATISTICAL ORBIT DETERMINATION Project Report Unscented kalman Filter Information.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
David Corne, and Nick Taylor, Heriot-Watt University - These slides and related resources:
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Satellite Tracking Example of SNC and DMC ASEN.
Discriminant Function Analysis Basics Psy524 Andrew Ainsworth.
Computer vision: models, learning and inference Chapter 19 Temporal models.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
Linear Regression Andy Jacobson July 2006 Statistical Anecdotes: Do hospitals make you sick? Student’s story Etymology of “regression”
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Modern Navigation Thomas Herring
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Factor Analysis Psy 524 Ainsworth. Assumptions Assumes reliable correlations Highly affected by missing data, outlying cases and truncated data Data screening.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2014 Professor Brandon A. Jones Lecture 26: Singular Value Decomposition.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
Modern Navigation Thomas Herring MW 11:00-12:30 Room
Processing Sequential Sensor Data The “John Krumm perspective” Thomas Plötz November 29 th, 2011.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
An Introduction to Kalman Filtering by Arthur Pece
Principles of the Global Positioning System Lecture 12 Prof. Thomas Herring Room ;
Principle Component Analysis and its use in MA clustering Lecture 12.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
Principal Component Analysis (PCA)
Error Modeling Thomas Herring Room ;
1 SVY 207: Lecture 6 Point Positioning –By now you should understand: F How receiver knows GPS satellite coordinates F How receiver produces pseudoranges.
Principles of the Global Positioning System Lecture 09 Prof. Thomas Herring Room A;
Principles of the Global Positioning System Lecture 08 Prof. Thomas Herring Room ;
SVY207: Lecture 10 Computation of Relative Position from Carrier Phase u Observation Equation u Linear dependence of observations u Baseline solution –Weighted.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2015 Professor Brandon A. Jones Lecture 22: Further Discussions of the.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Computacion Inteligente Least-Square Methods for System Identification.
University of Colorado Boulder ASEN 5070: Statistical Orbit Determination I Fall 2015 Professor Brandon A. Jones Lecture 19: Examples with the Batch Processor.
CSC2535: Computation in Neural Networks Lecture 7: Independent Components Analysis Geoffrey Hinton.
R. Kass/Sp07P416/Lecture 71 More on Least Squares Fit (LSQF) In Lec 5, we discussed how we can fit our data points to a linear function (straight line)
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
STATISTICAL ORBIT DETERMINATION Kalman (sequential) filter
STATISTICAL ORBIT DETERMINATION Coordinate Systems and Time Kalman Filtering ASEN 5070 LECTURE 21 10/16/09.
ASEN 5070: Statistical Orbit Determination I Fall 2014
Kalman Filtering: Control with Limited/Noisy Measurements
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
OVERVIEW OF LINEAR MODELS
Simple Linear Regression
OVERVIEW OF LINEAR MODELS
Principles of the Global Positioning System Lecture 11
Principles of the Global Positioning System Lecture 13
Kalman Filter: Bayes Interpretation
Probabilistic Surrogate Models
Presentation transcript:

Principles of the Global Positioning System Lecture 11 Prof. Thomas Herring Room A;

03/13/ Lec 112 Statistical approach to estimation Summary –Look at estimation from statistical point of view –Propagation of covariance matrices –Sequential estimation

03/13/ Lec 113 Statistical approach to estimation Examine the multivariate Gaussian distribution: By minimizing the argument of the exponential in the probability density function, we maximize the likelihood of the estimates (MLE). This is just weighted least squares where the weight matrix is chosen to be the inverse of the covariance matrix of data noise

03/13/ Lec 114 Data covariance matrix If we use the inverse of the covariance matrix of the noise in the data, we obtain a MLE if data noise is Gaussian distribution. How do you obtain data covariance matrix? Difficult question to answer completely Issues to be considered: –Thermal noise in receiver gives on component –Multipath could be treated as a noise-like quantity –Signal-to-noise ratio of measurements allows an estimate of the noise (discussed later in course). –In-complete mathematical model of observables can sometimes be treated as noise-like. –Gain of GPS antenna will generate lower SNR at low elevation angles

03/13/ Lec 115 Data covariance matrix In practice in GPS (as well as many other fields), the data covariance matrix is somewhat arbitrarily chosen. Largest problem is temporal correlations in the measurements. Typical GPS data set size for 24- hours of data at 30 second sampling is 8x2880=23000 phase measurements. Since the inverse of the covariance matrix is required, fully accounting for correlations requires the inverse of 23000x23000 matrix. To store the matrix would require, 4Gbytes of memory Even if original covariance matrix is banded (ie., correlations over a time short compared to 24-hours), the inverse of banded matrix is usually a full matrix

03/13/ Lec 116 Data covariance matrix Methods on handling temporal correlations: –If measurements correlated over say 5-minute period, then use samples every 5-minutes (JPL method) –Use full rate data, but artificially inflate the noise on each measurement so that equivalent to say 5-minute sampling (ie., sqrt(10) higher noise on the 30-second sampled values (GAMIT method) –When looking a GPS results, always check the data noise assumptions (discussed more near end of course). Assuming a valid data noise model can be developed, what can we say about noise in parameter estimates?

03/13/ Lec 117 Propagation of covariances Given a data noise covariance matrix, the characteristics of expected values can be used to determine the covariance matrix of any linear combination of the measurements.

03/13/ Lec 118 Propagation of covariance Propagation of covariance can be used for any linear operator applied to random variables whose covariance matrix is already known. Specific examples: –Covariance matrix of parameter estimates from least squares –Covariance matrix for post-fit residuals from least squares –Covariance matrix of derived quantities such as latitude, longitude and height from XYZ coordinate estimates.

03/13/ Lec 119 Covariance matrix of parameter estimates Propagation of covariance can be applied to the weighted least squares problem: Notice that the covariance matrix of parameter estimates is a natural output of the estimator if A T V -1 A is inverted (does not need to be)

03/13/ Lec 1110 Covariance matrix of estimated parameters Notice that for the rigorous estimation, the inverse of the data covariance is needed (time consuming if non-diagonal) To compute to parameter estimate covariance, only the covariance matrix of the data is needed (not the inverse) In some cases, a non-rigorous inverse can be done with say a diagonal covariance matrix, but the parameter covariance matrix is rigorously computed using the full covariance matrix. This is a non-MLE but the covariance matrix of the parameters should be correct (just not the best estimates that can found). This techniques could be used if storage of the full covariance matrix is possible, but inversion of the matrix is not because it would take too long or inverse can not be performed in place.

03/13/ Lec 1111 Covariance matrix of post-fit residuals Post-fit residuals are the differences between the observations and the values computed from the estimated parameters Because some of the noise in the data are absorbed into the parameter estimates, in general, the post-fit residuals are not the same as the errors in the data. In some cases, they can be considerably smaller. The covariance matrix of the post-fit residuals can be computed using propagation of covariances.

03/13/ Lec 1112 Covariance matrix of post-fit residuals This can be computed using propagation on covariances: e is the vector of true errors, and v is vector of residuals

03/13/ Lec 1113 Post-fit residuals Notice that we can compute the compute the covariance matrix of the post-fit residuals (a large matrix in generate) Eqn 1 on previous slide gives an equation of the form v=Be; why can we not compute the actual errors with e=B -1 v? B is a singular matrix which has no unique inverse (there is in fact one inverse which would generate the true errors) Note: In this case, singularity does not mean that there is no inverse, it means there are an infinite number of inverses.

03/13/ Lec 1114 Example Consider the case shown below: When a rate of change is estimated, the slope estimate will absorb error in the last data point particularly as  t increases. (Try this case yourself)

03/13/ Lec 1115 Covariance of derived quantities Propagation of covariances can be used to determine the covariance of derived quantities. Example latitude, longitude and radius.  is co-latitude, is longitude, R is radius.  N,  E and  U are north, east and radial changes (all in distance units).

03/13/ Lec 1116 Estimation in parts/Sequential estimation A very powerful method for handling large data sets, takes advantage of the structure of the data covariance matrix if parts of it are uncorrelated (or assumed to be uncorrelated).

03/13/ Lec 1117 Sequential estimation Since the blocks of the data covariance matrix can be separately inverted, the blocks of the estimation (A T V -1 A) can be formed separately can combined later. Also since the parameters to be estimated can be often divided into those that effect all data (such as station coordinates) and those that effect data a one time or over a limited period of time (clocks and atmospheric delays) it is possible to separate these estimations (shown next page).

03/13/ Lec 1118 Sequential estimation Sequential estimation with division of global and local parameters. V is covariance matrix of new data (uncorrelated with priori parameter estimates), V xg is covariance matrix of prior parameter estimates with estimates x g and x l are local parameter estimates, x g + are new global parameter estimates.

03/13/ Lec 1119 Sequential estimation As each block of data is processed, the local parameters, x l, can be dropped and the covariance matrix of the global parameters x g passed to the next estimation stage. Total size of adjustment is at maximum the number of global parameters plus local parameters needed for the data being processed at the moment, rather than all of the local parameters.

03/13/ Lec 1120 Summary We examined the way covariance matrices and be manipulated Estimation from a statistical point of view Sequential estimation. Next class continue with sequential estimation in terms of Kalman Filtering.