Variational Quality Control

Slides:



Advertisements
Similar presentations
Variational data assimilation and forecast error statistics
Advertisements

Assumptions underlying regression analysis
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Logistic Regression Psy 524 Ainsworth.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Correlation and regression
DA/SAT Training Course, March 2006 Variational Quality Control Erik Andersson Room: 302 Extension: 2627
ECMWF Slide 1 How does 4D-Var handle Nonlinearity and non-Gaussianity? Mike Fisher Acknowledgements: Christina Tavolato, Elias Holm, Lars Isaksen, Tavolato,
Lecture II-2: Probability Review
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Dee: Practice of Quality ControlNCAR Summer Colloquium Practice of Quality Control Dick Dee Global Modeling and Assimilation Office NASA Goddard.
Probability theory 2 Tron Anders Moger September 13th 2006.
Review of Statistical Models and Linear Regression Concepts STAT E-150 Statistical Methods.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Simple Linear Regression Analysis Chapter 13.
Lecture 8: Measurement Errors 1. Objectives List some sources of measurement errors. Classify measurement errors into systematic and random errors. Study.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Chapter 7: Hypothesis Testing. Learning Objectives Describe the process of hypothesis testing Correctly state hypotheses Distinguish between one-tailed.
Yandell – Econ 216 Chap 15-1 Chapter 15 Multiple Regression Model Building.
Data Modeling Patrice Koehl Department of Biological Sciences
Modeling and Simulation CS 313
Lecture Slides Elementary Statistics Twelfth Edition
Chapter 13 Simple Linear Regression
ECMWF/EUMETSAT NWP-SAF Satellite data assimilation Training Course
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Chapter 4 Basic Estimation Techniques
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Deep Feedforward Networks
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
BAE 5333 Applied Water Resources Statistics
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Modeling and Simulation CS 313
REGRESSION DIAGNOSTIC II: HETEROSCEDASTICITY
Recent Developments in the CP Violating TGC Analysis
Statistics in MSmcDESPOT
The Maximum Likelihood Method
Elementary Statistics
Latent Variables, Mixture Models and EM
Lecture Slides Elementary Statistics Thirteenth Edition
I271B Quantitative Methods
Statistical Methods For Engineers
CHAPTER 29: Multiple Regression*
6-1 Introduction To Empirical Models
Introduction to Econometrics, 5th edition
J.-F. Pâris University of Houston
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Undergraduated Econometrics
10701 / Machine Learning Today: - Cross validation,
Error statistics in data assimilation
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
EE513 Audio Signals and Systems
5.2 Least-Squares Fit to a Straight Line
M248: Analyzing data Block D UNIT D2 Regression.
Computing and Statistical Data Analysis / Stat 7
Product moment correlation
Parametric Methods Berlin Chen, 2005 References:
Diagnostics and Remedial Measures
Modeling with the normal distribution
Sarah Dance DARC/University of Reading
Model Adequacy Checking
Testing Causal Hypotheses
Probabilistic Surrogate Models
Presentation transcript:

Variational Quality Control Elias Holm Data Assimilation Section, ECMWF Acknowledgements to Lars Isaksen, Christina Tavolato and Erik Andersson, ECMWF DA/SAT Training Course

Introduction Assuming Gaussian statistics, the maximum likelihood solution to the linear estimation problem results in observation analysis weights (w) that are independent of the observed value. Outliers will be given the same weight as good data, potentially corrupting the analysis DA/SAT Training Course

QC-rejection or good data? Even good-quality data show significant deviations from the pure Gaussian form Actual distribution Gaussian “Tail” QC-rejection or good data? K obs-bg departures The real data distribution has fatter tails than the Gaussian Aircraft temperature observations shown here DA/SAT Training Course

The Normal observation cost function Jo (1) The general expression for the observation cost function is based on the probability density function (the pdf) of the observation error distribution (see Lorenc 1986): arbitrary constant, chosen such that Jo=0 when y=Hx p is the probability density function of observation error DA/SAT Training Course

The Normal observation cost function Jo (2) When for p we assume the normal (Gaussian) distribution (N): we obtain the expression Normalized departure y: observation x: represents the model/analysis variables H: observation operators σo: observation error standard deviation In VarQC a non-Gaussian pdf will be used, resulting in a non-quadratic expression for Jo. DA/SAT Training Course

Accounting for non-Gaussian effects In an attempt to better describe the tails of the observed distributions, Ingleby and Lorenc (1993) suggested a modified pdf (probability density function), written as a sum of two distinct distributions: Normal distribution (pdf), as appropriate for ‘good’ data pdf for data affected by gross errors A is the prior probability of gross error DA/SAT Training Course

Gross errors of that type occur occasionally… Positive observed temperatures (˚C) reported with wrong sign. (Chinese aircraft data 1-21 May 2007) Innovation Statistics Sample=429,000 All data Rejected data DA/SAT Training Course

Variational quality control Thus, a pdf for the data affected by gross errors (pG) needs to be specified. Several different forms could be considered. In the ECMWF 1998-2009 implementation (Andersson and Järvinen 1999, QJRMS) a flat distribution was chosen. 2d is the width of the distribution The consequence of this choice will become clear in the following DA/SAT Training Course

VarQC formulation Inserting pQC for p in the expression Jo=-ln p + const, we obtain: We can see how the presence of γ modifies the normal cost function and its gradient DA/SAT Training Course

Probability of gross error The term modifying the gradient (on the previous slide) can be shown to be equal to: the a-posteriori probability of gross error P, given x and assuming that Hx is correct (see Ingleby and Lorenc 1993) Furthermore, we can define a VarQC weight W: It is the factor by which the gradient’s magnitude is reduced. Data which are found likely to be incorrect (P≈1) are given reduced weight in the analysis. Data which are found likely to be correct (P ≈ 0) are given the weight they would have had using purely Gaussian observation error pdf. DA/SAT Training Course

Gaussian + flat PDF Sum of 2 Gaussians Gradient Gradient QC Weight DA/SAT Training Course

The application of VarQC is always in terms of the observed quantity. In the case of many observations, all with uncorrelated errors, JoQC is computed as a sum (over the observations i) of independent cost function contributions: The global set of observational data includes a variety of observed quantities, as used by the variational scheme through their respective observation operators. All are quality controlled together, as part of the main 4D-Var estimation. The application of VarQC is always in terms of the observed quantity. DA/SAT Training Course

Tuning the rejection limit The histogram on the left has been transformed (right) such that the Gaussian part appears as a pair of straight lines forming a ‘V’ at zero. The slope of the lines gives the Std deviation of the Gaussian. The rejection limit can be chosen to be where the actual distribution is some distance away from the ‘V’ - around 6 to 7 K in this case, would be appropriate. DA/SAT Training Course

Transforming the Gaussian pdf y: observation x: represents the model/analysis variables H: observation operators σo: observation error standard deviation DA/SAT Training Course

Tuning example BgQC too tough BgQC and VarQC correctly tuned The shading reflects the value of P, the probability of gross error DA/SAT Training Course

Tropical Cyclone example Observations of intense and small-scale features may be rejected although the measurements are correct. The problem occurs when the resolution of the analysis system (as determined by the B-matrix and model resolution) is insufficient. DA/SAT Training Course

Huber-norm an alternative A compromise between the l2 and l1 norms Robust method: a few erroneous observations does not ruin analysis Adds some weight on observations with large departures A set of observations with consistent large departures will influence the analysis Concave cost function Gaussian Huber norm (New operational) Gaussian + flat (old operational) DA/SAT Training Course

Huber norm variational quality control The pdf for the Huber norm is: Equivalent to L1 metric far from x, L2 metric close to x. With this pdf, observations far from x are given less weight than observations close to x, but can still influence the analysis. Many observations have errors that are well described by the Huber norm. DA/SAT Training Course

Transforming the Gaussian and exponential pdf y: observation x: represents the model/analysis variables H: observation operators σo: observation error standard deviation DA/SAT Training Course

Comparing observation weights: Huber-norm (red) versus Gaussian+flat (blue) More weight in the middle of the distribution More weight on the edges of the distribution More influence of data with large departures Weights: 0 – 25% 25% DA/SAT Training Course

Departure statistics for radiosonde temperatures is well described by a Huber-norm distribution Based on 18 months of data Feb 2006 – Sep 2007 Normalised fit of pdf to data Best Gaussian fit Best Huber norm fit All data Used data Data counts (log scaling) Normalized departures DA/SAT Training Course

METAR surface pressure data (Tropics) Blacklisting data may well contain gross errors Blacklisted data included After removing the blacklisted data the departures are well described by a Huber norm (black crosses & red line) Data counts (log scaling) What is left after removing blacklisted data Normalized departures DA/SAT Training Course

VarQC Summary VarQC is efficient quality control mechanism – all data types are quality controlled simultaneously as part of the 3D/4D-Var minimization. The implementation is very straight forward. VarQC does not replace the pre-analysis checks - the checks against the background for example. However, with Huber-norm these are relaxed significantly. DA/SAT Training Course