DA/SAT Training Course, March 2006 Variational Quality Control Erik Andersson Room: 302 Extension: 2627

Slides:



Advertisements
Similar presentations
Assumptions underlying regression analysis
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Linear Regression.
Logistic Regression Psy 524 Ainsworth.
Change Detection C. Stauffer and W.E.L. Grimson, “Learning patterns of activity using real time tracking,” IEEE Trans. On PAMI, 22(8): , Aug 2000.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Probability & Statistical Inference Lecture 9
STA305 week 31 Assessing Model Adequacy A number of assumptions were made about the model, and these need to be verified in order to use the model for.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
Visual Recognition Tutorial
Project  Now it is time to think about the project  It is a team work Each team will consist of 2 people  It is better to consider a project of your.
Statistical Background
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Correlation and Regression Analysis
Lecture II-2: Probability Review
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Relationships Among Variables
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Radial Basis Function Networks
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Dee: Practice of Quality ControlNCAR Summer Colloquium Practice of Quality Control Dick Dee Global Modeling and Assimilation Office NASA Goddard.
Model Inference and Averaging
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
ECE 8443 – Pattern Recognition LECTURE 10: HETEROSCEDASTIC LINEAR DISCRIMINANT ANALYSIS AND INDEPENDENT COMPONENT ANALYSIS Objectives: Generalization of.
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
Introduction to Inferential Statistics Statistical analyses are initially divided into: Descriptive Statistics or Inferential Statistics. Descriptive Statistics.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
Univariate Linear Regression Problem Model: Y=  0 +  1 X+  Test: H 0 : β 1 =0. Alternative: H 1 : β 1 >0. The distribution of Y is normal under both.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
1 2 nd Pre-Lab Quiz 3 rd Pre-Lab Quiz 4 th Pre-Lab Quiz.
LECTURE 3: ANALYSIS OF EXPERIMENTAL DATA
Chapter 2 Statistical Background. 2.3 Random Variables and Probability Distributions A variable X is said to be a random variable (rv) if for every real.
Model Selection and Validation. Model-Building Process 1. Data collection and preparation 2. Reduction of explanatory or predictor variables (for exploratory.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 12: Advanced Discriminant Analysis Objectives:
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 04: GAUSSIAN CLASSIFIERS Objectives: Whitening.
P Values - part 2 Samples & Populations Robin Beaumont 2011 With much help from Professor Chris Wilds material University of Auckland.
Chapter 20 Statistical Considerations Lecture Slides The McGraw-Hill Companies © 2012.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
- 1 - Preliminaries Multivariate normal model (section 3.6, Gelman) –For a multi-parameter vector y, multivariate normal distribution is where  is covariance.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Computacion Inteligente Least-Square Methods for System Identification.
Variational Quality Control
Deep Feedforward Networks
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Lecture Slides Elementary Statistics Thirteenth Edition
CHAPTER 29: Multiple Regression*
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Undergraduated Econometrics
EE513 Audio Signals and Systems
The Examination of Residuals
Multivariate Methods Berlin Chen
Multivariate Methods Berlin Chen, 2005 References:
Presentation transcript:

DA/SAT Training Course, March 2006 Variational Quality Control Erik Andersson Room: 302 Extension: 2627

DA/SAT Training Course, March 2006 Introduction VarQC formalism Rejection limits and tuning Minimisation aspects Examples Summary Plan of Lecture

DA/SAT Training Course, March 2006 Introduction (1) Assuming Gaussian statistics, the maximum likelihood solution to the linear estimation problem results in observation analysis weights (w) that are independent of the observed value. Outliers will be given the same weight as good data, potentially corrupting the analysis

DA/SAT Training Course, March 2006 Introduction (2) Real histogram distributions of departures (y-Hx b ) can show significant deviations from the pure Gaussian form. They often reveal a more frequent occurrence of large departures than expected from the corresponding Gaussian (normal) distribution with the same mean and standard deviation - showing as wide “Tails”. Observed departure from the background “Tail” QC-rejection? Actual distribution Gaussian K Airep temperature observations

DA/SAT Training Course, March 2006 The Normal Jo (1) The normal observation cost function Jo(x) has a quadratic form, which is consistent with the assumption that the errors are Gaussian in nature. If observations errors are uncorrelated then this simplifies to: y: array of observations of length=N x: represents the model/analysis variables H: observation operators R: observation error covariance matrix σ o : observation error standard deviation Normalized departure

DA/SAT Training Course, March 2006 The Normal Jo (2) The general expression for the observation cost function is based on the probability density function (the pdf) of the observation error distribution (see Lorenc 1986): p is the probability density function of observation error arbitrary constant, chosen such that Jo=0 when y=Hx

DA/SAT Training Course, March 2006 The Normal Jo (3) When for p we insert the normal (Gaussian) distribution (N): we obtain the usual expression In VarQC a non-Gaussian pdf will be used, resulting in a non-quadratic expression for Jo.

DA/SAT Training Course, March 2006 VarQC formulation (1) In an attempt to better describe the tails of the observed distributions, Ingleby and Lorenc (1993) suggested a modified pdf, written as a sum of two distinct distributions: A is the prior probability of gross error Normal distribution (pdf), as appropriate for ‘good’ data pdf for data affected by gross errors

DA/SAT Training Course, March 2006 VarQC formulation (2) Thus, a pdf for the data affected by gross errors (p G ) needs to be specified. Several different forms could be considered. In the ECMWF implementation (Andersson and Järvinen 1999, QJRMS) a flat distribution was chosen. The reasons for this choice and its consequences will become clear in the following D is the width of the distribution D is here written as a multiple d of the observation error, either side of zero.

DA/SAT Training Course, March 2006 VarQC formulation (3) Inserting p QC for p in the expression Jo=-ln p + const, we obtain: We can see how the presence of γ modifies the normal cost function and its gradient

DA/SAT Training Course, March 2006 Probability of gross error The term modifying the gradient (on the previous slide) can be shown to be equal to: the a-posteriori probability of gross error P, given x and assuming that Hx is correct (see Ingleby and Lorenc 1993) Furthermore, we can define a VarQC weight W: It is the factor by which the gradient’s magnitude is reduced. Data which are found likely to be incorrect (P≈1) are given reduced weight in the analysis. Data which are found likely to be correct (P ≈ 0) are given the weight they would have had using purely Gaussian observation error pdf.

DA/SAT Training Course, March 2006 Illustrations flat p QC (left), wide Gaussian p QC (right) Gradient QC Weight

DA/SAT Training Course, March 2006 Application In the case of many observations, all with uncorrelated errors, Jo QC is computed as a sum (over the observations i) of independent cost function contributions: The global set of observational data includes a variety of observed quantities, as used by the variational scheme through their respective observation operators. All are quality controlled together, as part of the main 4D-Var estimation. The application of VarQC is always in terms of the observed quantity.

DA/SAT Training Course, March 2006 Minimisation VarQC requires a good ‘preliminary analysis’. Otherwise incorrect QC-decisions will occur. In operations VarQC is therefore switched on after 40 iterations. (Now 40)

DA/SAT Training Course, March 2006 Multiple minima The probability of gross error depends on the size of the departure. When the probability for rejection is close to 50/50 this will be reflected in the cost function as two distinct minima of roughly equal magnitude.

DA/SAT Training Course, March 2006 Rejection limits VarQC does not require the specification of threshold values at which rejections occur - so called rejection limits. Rejections occur gradually. If, for example, we classify as rejected those data that have P>0.75, then we can obtain an analytical relationship for the effective rejection limit as a function of the two VarQC input parameters A and d (or γ) only: The first implementation of VarQC was thereby tuned to roughly reproduce the rejections of the old scheme (OIQC).

DA/SAT Training Course, March 2006 Tuning the rejection limit The histogram on the left has been transformed (right) such that the Gaussian part appears as a pair of straight lines forming a ‘V’ at zero. The slope of the lines gives the Std deviation of the Gaussian. The rejection limit can be chosen to be where the actual distribution is some distance away from the ‘V’ - around 6 to 7 K in this case, would be appropriate.

DA/SAT Training Course, March 2006 Tuning example BgQC too tough BgQC and VarQC correctly tuned The shading reflects the value of P, the probability of gross error

DA/SAT Training Course, March 2006 Example (1) For given values of the VarQC input parameters (A and d), the QC result (i.e. P), is a function of the normalized departure (y-Hx)/σ o, only.

DA/SAT Training Course, March 2006 Example (2) VarQC checks all data and all data types simultaneously. In this Australian example the presence of aircraft data has led to the rejection of a PILOT wind.

DA/SAT Training Course, March 2006 Example (3) - a difficult one Observations of intense and small-scale features may be rejected although the measurements are correct. The problem occurs when the resolution of the analysis system (as determined by the B-matrix) is insufficient.

DA/SAT Training Course, March 2006 Summary VarQC provides a satisfactory and very efficient quality control mechanism - consistent with 3D/4D-Var. The implementation can be very straight forward. VarQC does not replace the pre-analysis checks - the checks against the background for example. All observational data from all data types are quality control simultaneously, as part of the general 3D/4D-Var minimisation. The setting of VarQC parameters needs regular revision. A good description of background errors is essential for effective, flow-dependent QC