Empirical Localization of Observation Impact in Ensemble Filters Jeff Anderson IMAGe/DAReS Thanks to Lili Lei, Tim Hoar, Kevin Raeder, Nancy Collins, Glen.

Slides:



Advertisements
Similar presentations
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Advertisements

Regression and correlation methods
Stat 112: Lecture 7 Notes Homework 2: Due next Thursday The Multiple Linear Regression model (Chapter 4.1) Inferences from multiple regression analysis.
How do model errors and localization approaches affects model parameter estimation Juan Ruiz, Takemasa Miyoshi and Masaru Kunii
Effects of model error on ensemble forecast using the EnKF Hiroshi Koyama 1 and Masahiro Watanabe 2 1 : Center for Climate System Research, University.
3.11 Adaptive Data Assimilation to Include Spatially Variable Observation Error Statistics Rod Frehlich University of Colorado, Boulder and RAL/NCAR Funded.
MACHINE LEARNING 9. Nonparametric Methods. Introduction Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 
Assimilating Sounding, Surface and Profiler Observations with a WRF-based EnKF for An MCV Case during BAMEX Zhiyong Meng & Fuqing Zhang Texas A&M University.
Simultaneous Estimation of Microphysical Parameters and State Variables with Radar data and EnSRF – OSS Experiments Mingjing Tong and Ming Xue School of.
ASSIMILATION of RADAR DATA at CONVECTIVE SCALES with the EnKF: PERFECT-MODEL EXPERIMENTS USING WRF / DART Altuğ Aksoy National Center for Atmospheric Research.
Some Innovative Applications and Approaches Using Nudging Four Dimensional Data Assimilation: Lili Lei and David R. Stauffer Dept. of Meteorology, Penn.
Brian Ancell, Cliff Mass, Gregory J. Hakim University of Washington
Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 Click to edit Master title style.
Advanced data assimilation methods with evolving forecast error covariance Four-dimensional variational analysis (4D-Var) Shu-Chih Yang (with EK)
Advanced data assimilation methods- EKF and EnKF Hong Li and Eugenia Kalnay University of Maryland July 2006.
REGRESSION AND CORRELATION
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Lecture II-2: Probability Review
Linear Regression/Correlation
Correlation & Regression
Oceanography 569 Oceanographic Data Analysis Laboratory Kathie Kelly Applied Physics Laboratory 515 Ben Hall IR Bldg class web site: faculty.washington.edu/kellyapl/classes/ocean569_.
6.1 What is Statistics? Definition: Statistics – science of collecting, analyzing, and interpreting data in such a way that the conclusions can be objectively.
Ensemble Data Assimilation and Uncertainty Quantification Jeffrey Anderson, Alicia Karspeck, Tim Hoar, Nancy Collins, Kevin Raeder, Steve Yeager National.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Observing Strategy and Observation Targeting for Tropical Cyclones Using Ensemble-Based Sensitivity Analysis and Data Assimilation Chen, Deng-Shun 3 Dec,
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
The Importance of Atmospheric Variability for Data Requirements, Data Assimilation, Forecast Errors, OSSEs and Verification Rod Frehlich and Robert Sharman.
Eidgenössisches Departement des Innern EDI Bundesamt für Meteorologie und Klimatologie MeteoSchweiz Statistical Characteristics of High- Resolution COSMO.
Ch4 Describing Relationships Between Variables. Pressure.
Regression Maarten Buis Outline Recap Estimation Goodness of Fit Goodness of Fit versus Effect Size transformation of variables and effect.
Federal Department of Home Affairs FDHA Federal Office of Meteorology and Climatology MeteoSwiss High-resolution data assimilation in COSMO: Status and.
Ch4 Describing Relationships Between Variables. Section 4.1: Fitting a Line by Least Squares Often we want to fit a straight line to data. For example.
Summary of introduced statistical terms and concepts mean Variance & standard deviation covariance & correlation Describes/measures average conditions.
Applications of optimal control and EnKF to Flow Simulation and Modeling Florida State University, February, 2005, Tallahassee, Florida The Maximum.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Sensitivity Analysis of Mesoscale Forecasts from Large Ensembles of Randomly and Non-Randomly Perturbed Model Runs William Martin November 10, 2005.
Development of an EnKF to estimate CO 2 fluxes from realistic distributions of X CO2 Liang Feng, Paul Palmer
11 Background Error Daryl T. Kleist* National Monsoon Mission Scoping Workshop IITM, Pune, India April 2011.
Pg 1 A Loosely Coupled Ocean-Atmosphere Ensemble Assimilation System. Tim Hoar, Nancy Collins, Kevin Raeder, Jeffrey Anderson, NCAR Institute for Math.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Data assimilation and forecasting the weather (!) Eugenia Kalnay and many friends University of Maryland.
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 6.
A kernel-density based ensemble filter applicable (?) to high-dimensional systems Tom Hamill NOAA Earth System Research Lab Physical Sciences Division.
Impacts of Running-In-Place on LETKF analyses and forecasts of the 24 May 2011 outbreak Impacts of Running-In-Place on LETKF analyses and forecasts of.
LWG, Destin (Fl) 27/1/2009 Observation representativeness error ECMWF model spectra Application to ADM sampling mode and Joint-OSSE.
Chapter 13 (Prototype Methods and Nearest-Neighbors )
Implementation and Testing of 3DEnVAR and 4DEnVAR Algorithms within the ARPS Data Assimilation Framework Chengsi Liu, Ming Xue, and Rong Kong Center for.
A Random Subgrouping Scheme for Ensemble Kalman Filters Yun Liu Dept. of Atmospheric and Oceanic Science, University of Maryland Atmospheric and oceanic.
Pg 1 SAMSI UQ Workshop; Sep Ensemble Data Assimilation and Uncertainty Quantification Jeff Anderson National Center for Atmospheric Research.
Future Directions in Ensemble DA for Hurricane Prediction Applications Jeff Anderson: NCAR Ryan Torn: SUNY Albany Thanks to Chris Snyder, Pavel Sakov The.
The application of ensemble Kalman filter in adaptive observation and information content estimation studies Junjie Liu and Eugenia Kalnay July 13th, 2007.
École Doctorale des Sciences de l'Environnement d’ Î le-de-France Année Modélisation Numérique de l’Écoulement Atmosphérique et Assimilation.
This represents the most probable value of the measured variable. The more readings you take, the more accurate result you will get.
June 20, 2005Workshop on Chemical data assimilation and data needs Data Assimilation Methods Experience from operational meteorological assimilation John.
The simple linear regression model and parameter estimation
Jeffrey Anderson, NCAR Data Assimilation Research Section
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Jeffrey Anderson, NCAR Data Assimilation Research Section
Jeff Anderson, NCAR Data Assimilation Research Section
A New Scheme for Chaotic-Attractor-Theory Oriented Data Assimilation
Jeff Anderson, NCAR Data Assimilation Research Section
Observation Informed Generalized Hybrid Error Covariance Models
Statistical Methods For Engineers
No notecard for this quiz!!
FSOI adapted for used with 4D-EnVar
Error statistics in data assimilation
Impact of Assimilating AMSU-A Radiances on forecasts of 2008 Atlantic TCs Initialized with a limited-area EnKF Zhiquan Liu, Craig Schwartz, Chris Snyder,
Project Team: Mark Buehner Cecilien Charette Bin He Peter Houtekamer
Sarah Dance DARC/University of Reading
Presentation transcript:

Empirical Localization of Observation Impact in Ensemble Filters Jeff Anderson IMAGe/DAReS Thanks to Lili Lei, Tim Hoar, Kevin Raeder, Nancy Collins, Glen Romine, Chris Snyder, Doug Nychka 5th EnKF Workshop, 23 May 20121

For an observation y and state variable x; Increments for N ensemble samples of x are: Where is a sample regression coefficient, and is a localization. Traditionally, but here there is no upper bound. 5th EnKF Workshop, 23 May Definition of Localization

5th EnKF Workshop, 23 May Empirical Localization Have output from an OSSE. Know prior ensemble and truth for each state variable.

5th EnKF Workshop, 23 May Empirical Localization Have output from an OSSE. Know prior ensemble and truth for each state variable. Can get truth & prior ensemble for any potential observations.

5th EnKF Workshop, 23 May Empirical Localization Estimate localization for set of observations and subset of state variables. e.g. state variables at various horizontal distances from observations.

5th EnKF Workshop, 23 May Empirical Localization Example: how to localize impact of temperature observations (4 shown) on a U state variable that is between 600 and 800 km distant.

5th EnKF Workshop, 23 May Empirical Localization Given observational error variance, can compute expected ensemble mean increment for state. Plot this vs prior state truth - ensemble mean.

5th EnKF Workshop, 23 May Empirical Localization Do this for all state variables in subset.

5th EnKF Workshop, 23 May Empirical Localization Do this for all state variables in subset.

5th EnKF Workshop, 23 May Empirical Localization Do this for all state variables in subset.

5th EnKF Workshop, 23 May Empirical Localization Find a least squares fit. Slope is. Least squares minimizes: Same as minimizing Posterior mean

5th EnKF Workshop, 23 May Empirical Localization Define set of all pairs (y, x) of potential observations and state variable instances in an OSSE. (A state variable instance is defined by type, location and time). Choose subsets of this set.

5th EnKF Workshop, 23 May Empirical Localization Find that minimizes the RMS difference between the posterior ensemble mean for x and the true value over this subset. This can be computed from the output of the OSSE. Can then use this localization in a new OSSE for all (y, x) in the subset. Call the values of localization for all subsets an Empirical Localization Function (ELF).

5th EnKF Workshop, 23 May Lorenz Variable Examples Assume all observations are located at a model grid point. (Easier but not necessary). Define 40 subsets of (y, x) pairs: x is 20 to the left, 19 to the left, … 1 to the left, colocated, 1 to the right, …, 19 to the right of y.

5th EnKF Workshop, 23 May Computing ELFs Start with a climatological ensemble. Do set of 6000-step OSSEs. (only use last 5000 steps). First has no localization. Compute ELF from each. Use ELF for next OSSE. ELF1 ELF2 ELF3 ELF4 ELF5 No Localization

5th EnKF Workshop, 23 May Evaluation Experiments Start with a climatological ensemble. Do 110,000 step assimilation, discard first 10,000 steps. Adaptive inflation with 0.1 inflation standard deviation. Many fixed Gaspari-Cohn localizations tested for each case. Also five ELFs (or should it be ELVEs?).

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Identity observations. Error variance 16. Assimilate every standard model timestep.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. N=20 Gaspari Cohn (GC) function with smallest time mean prior RMSE.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. N=20 first ELF is negative for many distances, but minimum localization is 0 when this ELF is used.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Subsequent N=20 ELFs are less negative, smoother, closer to best GC.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Subsequent N=20 ELFs are less negative, smoother, closer to best GC.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Subsequent N=20 ELFs are less negative, smoother, closer to best GC.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Subsequent N=20 ELFs are less negative, smoother, closer to best GC.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. N=20, best GC has half-width 0.2, time mean RMSE of ~1.03.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. N=20, best GC has half-width 0.2, time mean RMSE of ~1.03. ELFs give RMSE nearly as small as this.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. N=20, best GC has half-width 0.2, time mean RMSE of ~1.03. ELFs give RMSE nearly as small as this.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Similar results for smaller ensemble, N=10. Note larger RMSE, narrower best GC half-width.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. Similar results for larger ensemble, N=40. Note smaller RMSE, wider best GC half-width.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. N=40 ELFs have smaller time mean RMSE than best GC.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. ELFs are nearly symmetric so can ignore negative distances.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. ELF for smaller ensemble is more compact.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. ELF for larger ensemble less compact, consistent with GC results.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. ELFs for even bigger ensembles are broader, but noisier at large distances.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. ELFs for even bigger ensembles are broader, but noisier at large distances.

5th EnKF Workshop, 23 May Case 1: Frequent low-quality obs. ELFs for even bigger ensembles are broader, but noisier at large distances.

5th EnKF Workshop, 23 May Case 2: Infrequent high-quality obs. Identity observations. Error variance 1. Assimilate every 12 th standard model timestep.

5th EnKF Workshop, 23 May Case 2: Infrequent high-quality obs. For N=10, all ELF cases have smaller RMSE than best GC.

5th EnKF Workshop, 23 May Case 2: Infrequent high-quality obs. For N=20, first ELF is worse than best GC; all others better. Best GC gets wider as ensemble size grows.

5th EnKF Workshop, 23 May Case 2: Infrequent high-quality obs. For N=40, all ELFs have smaller RMSE.

5th EnKF Workshop, 23 May Case 2: Infrequent high-quality obs. N=10 ELF is non-Gaussian. Has local minimum localization for distance 1.

5th EnKF Workshop, 23 May Case 2: Infrequent high-quality obs. N=40 ELF is broader; also has local minimum for distance 1. Need a non-gaussian ELF to possibly do better than GC.

5th EnKF Workshop, 23 May Case 3: Integral observations. Each observation is average of grid point plus its nearest 8 neighbors on both side; total of 17 points. (Something like a radiance observation.)

5th EnKF Workshop, 23 May Case 3: Integral observations. Each observation is average of grid point plus its nearest 8 neighbors on both side; total of 17 points. (Something like a radiance observation.) Error variance 1. Assimilate every standard model timestep. Very low information content: Assimilate 8 of these observations for each grid point. Total of 320 observations per assimilation time.

5th EnKF Workshop, 23 May Case 3: Integral observations. ELFs are not very Gaussian. No values close to 1, two peaks at distance +/- 7.

5th EnKF Workshop, 23 May Case 3: Integral observations. ELFs are not very Gaussian. Best GC is much larger near the observation location.

5th EnKF Workshop, 23 May Case 3: Integral observations. RMSE is a more complicated function of the GC half-width in this case.

5th EnKF Workshop, 23 May Case 3: Integral observations. ELFs all have significantly smaller time mean RMSE than best GC.

5th EnKF Workshop, 23 May Case 4: Frequent low-quality obs., imperfect model Identity observations. Error variance 16. Assimilate every standard model timestep. Truth has forcing F=8 (chaotic). Ensemble has forcing F=5 (not chaotic).

5th EnKF Workshop, 23 May Case 4: Frequent low-quality obs., imperfect model These are the localizations for the Case 1 perfect model.

5th EnKF Workshop, 23 May Case 4: Frequent low-quality obs., imperfect model Best GC is more compact for imperfect model case. Fifth ELF also more compact, but not as close to imperfect GC.

5th EnKF Workshop, 23 May How long an OSSE does this take? For large localization get good results with O(100) OSSE steps. Errors grow much more quickly for small localizations.

5th EnKF Workshop, 23 May Conclusions Can get estimates of good localization for any subset of observations and state variables from an OSSE. If good localizations are non-Gaussian do better than Gaspari Cohn. When Gaussian, can still be cheaper than tuning half-widths. Can this be applied to real geophysical models? How much could real applications be improved? Unclear… Can localization functions be separable in large models? Loc(time diff) * Loc(horizontal dist.) * Loc(vertical dist.) * Loc(obs_type, state_type)???

5th EnKF Workshop, 23 May Related Activities: Lili Lei Poster Testing ELFs in global climate model (CAM), and in WRF regional nested configuration. Some results look very similar to earlier sampling error correction methods. Next step, using ELFs in iterated OSSEs.

5th EnKF Workshop, 23 May Empirical Localization Without Knowing Truth Find that minimizes the RMSE between the posterior ensemble mean for x and observed value of x over the subset of (y, x) pairs. This can be computed from the output of an assimilation. Can then use this localization in a new assimilation for all (y, x) in the subset. BUT, can only compute for pairs of OBSERVED quantities. Can act as a way to calibrate OSSE results?

5th EnKF Workshop, 23 May Case 1 without knowing truth Identity observations. Error variance 16. Assimilate every standard model timestep. All state variables are observed, so no problem there.

5th EnKF Workshop, 23 May Case 1 without knowing truth Using real obs is much noisier for small localization values. Similar to using truth for larger localization values. Could be used to calibrate results from an OSSE for real assimilation use.