We think you have liked this presentation. If you wish to download it, please recommend it to your friends in any social system. Share buttons are a little bit lower. Thank you!
Presentation is loading. Please wait.
Published byAllison O'Grady
Modified over 2 years ago
Page 1© Crown copyright Operational Use of ECMWF products at the Met Office: Current practice, Verification and Ideas for the future Tim Hewson 17 th June 2005
Page 2© Crown copyright Contents Describe use of ECMWF output, in the Operations Centre at the Met Office (which provides guidance to the outfield), in 3 forecast categories: 1. Short range (up to 36 hours) 2. Medium range (36 hours to 6 days) 3. Trend period (6 to 10 days) 4. Some verification ideas (severe weather)
Page 3© Crown copyright 1. Short Range Relatively recent changes in ECMWF operational suite (2 operational runs per day, much reduced delay in forecast arrival time, etc) render ECMWF operational runs much more useful for short range forecasting than hitherto Data time (DT) difference c.f. Met Office operational runs now averages 9 hours (previously it was 21 hours) 9 hours is comparable to the short period lead time gain of ECMWF over Met Office (10 hours), implying high potential for adding value to deterministic forecasts using the consensus (poor mans ensemble) approach Though not necessarily part of the ECMWF remit if ECMWF model output can potentially improve our short range forecasts we will use it Some use also made of ensemble data (though much less than in the medium range – due to reduced reliability of severe weather probabilities)
Page 4© Crown copyright Operational Model Errors 1 5 days 10 RANK Best - EC UK FR.. NH rms MSLP error vs Lead Time Lead time gain = 10 hours short range | medium range | trend period |
Page 5© Crown copyright RAW MOD OBS Short Range Forecast Example RAW Cloud cover Ppn rate Ppn type Mslp
Page 6© Crown copyright What was it, specifically, about the good forecast that made it better? Helps highlight common model errors Verification of Modifications
Page 7© Crown copyright Scope for Improvement Forecasters are usually able to improve upon raw (Met Office) model output, using different models, knowledge of systematic errors, comparison with current trends Degree of improvement could potentially be increased by making more use of the high quality ECMWF operational run, which at present is under-utilised WISH LIST! – 3 hourly data, T+0 to T+48 Instantaneous total ppn rates, plus cloud cover and mslp (same format?!) Separate plots showing dynamic /convective rain and snow components 10m mean wind and likely gust strength Sub areas – parts of Europe ? Timely appearance on ECMWF web site is crucial (probably the most expedient route for making this data available)
Page 8© Crown copyright Severe Weather Table records Significant Errors, from perspective of hazardous weather, in Raw (Red) and Modified (Blue) forecasts (2.5 years of data) A key forecasting target is a reduction in the number of blue boxes This verification has highlighted warm air summer convection as one area where the forecaster is able to add little value, and where there can be major errors in the model Potentially ECMWF oper runs could assist in this area, via the wish list data? (convective characteristics?) Cold air convection is also a problem area, though one that the forecaster often addresses reasonably successfully. FGEW verification has highlighted this as a significant weakness in ECMWF output (snow in NElies, Feb/Mar 2005) Low pressure centres and surface wind gusts in strong gradient regions are another sig problem
Page 9© Crown copyright 2. Medium Range Use of ECMWF data - operational and ensemble runs - is more common in the medium range Timeliness is a key issue for practical applications; recently the ECMWF web site has been used more often, because of timely updating Issued forecast guidance has both deterministic and probabilistic components
Page 10© Crown copyright Deterministic Component Underlying Strategy: First ascertain the key meteorological feature(s), Then incorporate different model and ensemble solutions accordingly, weighting as appropriate (on screen or on paper) to arrive at a consensus solution
Page 11© Crown copyright Operational Model Errors 1 5 days 10 NH rms MSLP error vs Lead Time short range | medium range | trend period |
Page 12© Crown copyright Cold front is the main feature Consensus would move GM front south, possibly hinting at wave development, as GM underdoes waves, and as NCEP shows one. Simple Example to Illustrate
Page 13© Crown copyright Modifying a Selected Model Run Use Field Modification tool (devised by Eddy Carroll) AFTER a model run has finished Allows quick, interactive, dynamically consistent changes to be made to a set of 3-d fields from one model run, eg: Move low or front Deepen/fill low Introduce low or wave… Relies on modification vectors applied to PV distribution, followed by PV inversion with changed boundary conditions Equivalent translation vectors applied to ppn and RH; simple boundary layer model used for surface winds Temporal consistency achieved via time-linking (with decay parameter) Precipitation rate & type, winds etc can also be adjusted directly and time-linked
Page 14© Crown copyright Field Modification Example – moving a low with slight deepening
Page 15© Crown copyright Resulting fields (takes ~2 seconds)
Page 16© Crown copyright Initial Fields
Page 17© Crown copyright 500mb ht and mb Thickness - before
Page 18© Crown copyright 500mb ht and mb Thickness - after
Page 19© Crown copyright Objective Verification
Page 20© Crown copyright Subjective Verification Met Office Global model, ECMWF Oper and issued Modified forecasts are compared, primarily for mslp, but also for fronts and thickness, over NW Europe, using 12Z data times, for T+48,72, 96 and 120h The overall ranking is (best first): Modified >> ECMWF >> UKMO This is despite the objectively calculated lead time gain of ECMWF over UKMO (10-16 hours) being generally greater than it is for Modified over UKMO (0-12 hours) Objective schemes can penalise good forecasts of cyclogenesis (especially rms errors) A new more discriminating subjective verification scheme will be introduced this year
Page 21© Crown copyright Forecaster Impact Modification Time
Page 22© Crown copyright Probabilistic component Consists of a headline summary of probabilities of severe weather, in a number of categories (introduced relatively recently) Minima and maxima for sites are issued with upper and lower bounds. Probabilities are given for rainfall total exceedance. Starting point is calibrated ECMWF ensemble output. Alternative scenarios can be issued to highlight uncertainties Probabilistic components appear to be used far less by customers than the deterministic component. Somewhat disappointing – education required, but may take a long time.
Page 23© Crown copyright 3. Trend Forecast Consists of an issued worded forecast, with expected temperature anomalies, and considerable discussion of uncertainties, highlighting possible severe weather Based primarily on ECMWF data, but with some input from other models and ensembles, notably from NCEP Customer base for this forecast has been dwindling (perhaps they watch Swedish TV)
Page 24© Crown copyright Trend Period – Subjective Verification Based primarily on mslp, over NW Europe, ECMWF only Ensemble mean rated a little more useful than Oper run, but not by much Only 1 in 5 forecasts for days 6 and 7 were considered useful, and 1 in 10 forecasts for days 8, 9 and10 Scores have not changed a great deal over the years
Page 25© Crown copyright Extras for Wish List Meteograms to include overlapping 24-hour rainfall totals (but still in 6 hour blocks) Total cloud cover – is this altitude weighted, or is 8 oktas cirrus considered cloudy? Weighting would be preferable Postage stamps showing estimated surface gusts, with colour-shading for high values Cluster ensemble means for mslp, thickness, annotated with percentages of members
Page 26© Crown copyright 4. Verification – some ideas Conceptually there should be a deterministic limit for predicting a pre-defined meteorological event Simply defined this could be the point in lead time beyond which forecasts concerning that event are more likely, on average, to be wrong than right Defined in this way, this provides some guidance on when to shift the emphasis, in forecasts for particular events, towards the probabilistic For rare events at least, correct null forecasts – ie the majority - can be ignored as not relevant
Page 27© Crown copyright Verification ideas (contd) The deterministic limit for the event in question is then simply the lead time at which, over a suitably large forecast sample, hits equals the sum of misses and false alarms (or CSI = 0.5) misses + false alarms hits number deterministic limit lead time dc ba Fc Ob a/(b+c) = 1
Page 28© Crown copyright Event Examples (numbers are very crude estimates) Tornado within 2km radius (deterministic limit ~ 5 mins) Snow falling at a point (~5 hours) Rain falling at a point (~18 hours) Gale force gusts at a point (~6 hours) Gale force gusts within a UK county (~24 hours) Rainfall >15mm in 3 hours somewhere in a UK county (2 hours) Cyclonic surface pressure pattern at a point (~120 hours) Atmospheric front within 200km of a point (~60 hours) Day with maximum above 30C in London (~96 hours)
Page 29© Crown copyright Benefits Potential to provide a meaningful measure of what to expect from, and therefore what to put into, a forecast. Too many forecast elements are deterministic. It is something that the public, other customers (and auditors!) could potentially relate to The equivalent, from an idealised, reliable ensemble prediction system, would be the lead time at which the average probability, for hindcasts of observed past events, fell to 50% (?) As always extreme events would be more difficult to represent (though hindcasts from re-analyses are becoming increasingly tractable) Facility also to measure forecast improvements, compare systems, assess forecaster performance
Page 30© Crown copyright Summary The Met Office makes extensive use of ECMWF products for forecasts from T+48 to T+240, and is increasingly using the operational run as an input to short range forecasts Verification indicates that the forecaster is adding value in many areas, in part using the poor mans ensemble approach, though some weaknesses remain Various enhancements to ECMWF web-based output have been suggested Disappointingly, the customer base for probabilistic forecasts is currently limited More guidance on what we can and cannot forecast deterministically is required A new measure of deterministic limit has been tentatively proposed
Page 31© Crown copyright References Carroll, Meteorological Applications, 1997, for field modification description Carroll and Hewson, Weather and Forecasting, 2005 (out shortly) for ops centre practice and verification
Page 32© Crown copyright Accreditation WAFC World Area Forecast Centre
Page 1© Crown copyright Some Strengths and Weaknesses of ECMWF Forecasts for the UK Tim Hewson 15 th June 2006 Contributors include: Eleanor Crompton,
Page 1© Crown copyright 2005 Use of EPS at the Met Office Ken Mylne and Tim Legg.
© Crown copyright Met Office Scientific background and content of new gridded products Bob Lunnon, Aviation Outcomes Manager, Met Office WAFS Workshop.
Training Course 2009 – NWP-PR: The Seasonal Forecast System at ECMWF 1 The Seasonal Forecast System at ECMWF Tim Stockdale European Centre for Medium-Range.
Page 1© Crown copyright 2004 Presentation to ECMWF Forecast Product User Meeting 16th June 2005.
Work measurement Part II of Work Study. 2 Introduction Work measurement is the application of techniques designed to establish the time for a qualified.
IWTC VI – Costa Rica – November 2006 Summary Session – Jim Davidson Quantitative Forecasts of TC Landfall in relation to an Effective Warning System.
Page 1© Crown copyright 2004 ECMWF Forecast Products Users Meeting 15th June 2006.
1 by L Goel Professor & Head of Division of Power Engineering School of Electrical & Electronic Engineering Nanyang Technological University,
DTC Verification Workshop - Boulder April Inference Methods Ian Jolliffe Inference Methods Ian Jolliffe.
ECMWF Slide 1Met Op training course – Reading, March 2004 Forecast verification: probabilistic aspects Anna Ghelli, ECMWF.
AREP GAW Section 12 Air Quality Forecasting Tools.
Slide 1 Forecast Products User Meeting June 2006 Summary Forecast Products User Meeting: June 2006 Summary.
2 Welcome To Defect Management Training Objective: The objective of this course is to learn about standards that emphasize a best practice approach for.
Robin Hogan Ewan OConnor, Anthony Illingworth University of Reading, UK Clouds radar collaboration meeting 17 Nov 09 Ground based evaluation of cloud forecasts.
1 Psychological Practical (Year 2) PS2001 Introduction Dr. John Beech.
AREP GAW Section 13 Developing a Forecasting Program Understanding Users Needs Understanding the Processes that Control Air Quality Choosing Forecasting.
HB 1 Forecast Products Users'Meeting, June 2005 Users meeting Summary Performance of the Forecasting System (1) Main (deterministic) model -Outstanding.
The THORPEX Interactive Global Ensemble (TIGGE) Multi-model ensembles and Tropical cyclone forecasting Richard Swinbank, with thanks to the GIFS-TIGGE.
Direct Time study: Selecting and timing the job First step in time study is to select the job to be studied. There is always a reason why a particular.
Chapter 3: Supervised Learning. CS583, Bing Liu, UIC 2 Road Map Basic concepts Decision tree induction Evaluation of classifiers Rule induction Classification.
Seasonal forecasting: status and plans David Anderson Tim Stockdale, Magdalena Balmasda, Arthur Vidard, Alberto Troccoli, Paco Doblas-Reyes, Kristian Morgensen,
Chapter 6 Three Simple Classification Methods The Naïve Rule Naïve Bayes k-Nearest Neighbor 1.
ECMWF User Meeting / 1 Pertti Nurmi Juha Kilpinen Annakaisa Sarkanen ( Finnish Meteorological Institute ) Probabilistic Forecasts.
© Copyright QinetiQ limited 2007 QinetiQ Proprietary Human Aspects of NEC: Decision- Making, Organisation and Information Dr Andy Belyavin.
1 GREY BOX TESTING Web Apps & Networking Session 10 Boris Grinberg
'Trends, time series and forecasting Paul Fryers East Midlands KIT.
Using Learning Outcomes and Assessment Criteria Peter Noakes Department of Electronic Systems Engineering University of Essex.
AS-201 Weather Forecasting. This section discusses: 1.Various weather forecasting methods, their tools, and forecasting accuracy and skill 2.Images for.
Training Course 2009 – NWP-PR: Calibration of EPSs 1/42 Calibration of EPSs Renate Hagedorn European Centre for Medium-Range Weather Forecasts.
© 2016 SlidePlayer.com Inc. All rights reserved.