Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 A Forecast Evaluation Tool (FET) for CPC Operational Forecasts Edward O’Lenic Chief, Operations Branch, NOAA-NWS-CPC.

Similar presentations


Presentation on theme: "1 A Forecast Evaluation Tool (FET) for CPC Operational Forecasts Edward O’Lenic Chief, Operations Branch, NOAA-NWS-CPC."— Presentation transcript:

1 1 A Forecast Evaluation Tool (FET) for CPC Operational Forecasts Edward O’Lenic Chief, Operations Branch, NOAA-NWS-CPC

2 2 Outline CPC operational products We have discovered users Users are diverse We need to know how users use products We need a new service paradigm FET addresses these needs Tour of FET Current status Implications

3 3 Background People and organizations need planning tools. For decades, climate forecasts have been sought for use in planning. The obvious signs of climate change have made this desire more urgent. The uncertain nature of climate forecasts can lead the uninformed to make poor choices. Users need an “honest broker” they can trust.

4 4 Current Operational Capabilities CPC products, include extreme events for days 3-14, extended-range forecasts of T, P for week 2, 1-month and 3-month outlooks, and ENSO forecasts. These products have a scientific basis, have skill, and therefore ought to be useful, but - We have a less-than adequate understanding of user requirements, or actual practical use. CPC has neither the staff, nor the expertise to effectively pursue understanding these things.

5 5 Challenges from Users Western Governor’s Association (Jones, 2007): - More accurate, finer-resolution long range forecasts - Continued and expanded funding for data collection, monitoring and prediction - Partnerships with federal and state climatologists, RCCs, agricultural extension services, resource management agencies, federal, state and local governments. USDA ARS Grazinglands Research Laboratory (Schneider, 2002) - Fewer “EC” forecasts - Better correspondence between F probability and O frequency - Forecast more useful than climatology - Forecasts of impacts, not meteorological variables

6 6 Bridging Research, Operations, Users Assumed User Needs ? GOVT. PROVIDERS Use it if you can Basic data and Forecasts Basic Research Know WHO stakeholders are and HOW they USE climate products, (relationships), LEARN from stakeholders WHAT we need to provide (iterative refinement of requirements, relationships), Implement USER-, and SCIENCE- VETTED products, (iterative refinement of requirements, relationships), Operationally support evolving user and producer REQUIREMENTS (extension function, relationships, operations, research)

7 7 Bridging Research, Operations, Users Intermediary Applications Products CTB-RISA/PRIVATE/RCC/SC-SBIR, NIDIS Decision-Support Development ? GOV PROVIDERS (CPC/EMC) NCEP/NCDC/USGS CTB SUPPOPRT Iterative Use refinement Transfer User-Vetted products PROVIDERS Iterative Technical Refinement Use it if you can Basic data, Forecasts RESEARCH, MODELING (dyn, stat), OBSERVATIONS GOV OPS PRIVATE OPS Transfer User-Vetted products O2R: Model Test Facility R2O : CTB users Research, Modeling, Obs O2R R2O REQUIREMENTS Assumed User Needs O2R R2O

8 8 Доверяй, но проверяй

9 9 Trust, but verify

10 10 Лроверяй, и Доверяй, Verify, and Trust

11 11 Trust is at the core of a successful product suite. Transparent verification is one way to secure trust. Лроверяй, и Доверяй, Verify, and Trust

12 12 Trust is at the core of a successful product suite. Transparent verification is one way to secure trust. VERIFICATION: A MEASURE OF FORECAST QUALITY, SKILL AND VALUE Лроверяй, и Доверяй, Verify, and Trust

13 13 Types of Verification Accuracy – e.g., AC, error, #correct, rmse, mae, etc… Skill – e.g., HSS, RPSS, Brier SS Bias – forecast too high/low? Resolution – how well are different events discriminated? Sharpness – able to predict extreme events?

14 14 How to Proceed? Over the last decade, Dr. Holly C. Hartmann, and programmers Ellen Lay and Damian Hammond, of CLIMAS, have developed an interactive, on-line “Forecast Evaluation Tool” (FET) which allows users to evaluate the meaning and skill of CPC 3-Month Outlooks of temperature and precipitation. CPC proposes to: 1)Make the FET CPC’s outlet to users for forecast skill information, 2)Become a partner with CLIMAS and others to make the FET a community resource. 3)Expand the capabilities of FET

15 15 Forecast Evaluation Tool: Example of a Means to Address Gaps What FET provides: User-centric forecast evaluation and data access and display capability. Leveraging of community software development capabilities. Opportunity to DISCOVER and collect user requirements.

16 16 CPC Temperature and Precipitation Outlooks Probability of three categories Bottom, middle, top 10 years, 1971-2000 Maps show probability of likeliest category Each category has a value at every point Probabilities sum to 100% If middle favored, borrow from extremes “EC” means pr(b,n,a)=33.33, 33.33, 33.33

17 17 CPC Temperature and Precipitation Outlooks Probability of three categories Bottom, middle, top 10 years, 1971-2000 Maps show probability of likeliest category Each category has a value at every point Probabilities sum to 100% If middle favored, borrow from extremes “EC” means pr(b,n,a)=33.33, 33.33, 33.33 Pr(b,n,a)=40, 33.33, 26.67 Pr(b,n,a)= 16.67, 33.33, 50

18 18 Downscaling: Probabilities to Temperatures

19 19 Downscaling: Probabilities to Amounts

20 20 Recent CPC Skill Improvements HSS=% improvement the forecast makes over random forecasts. More non-EC forecasts, and a large HSS are GOOD. Non-EC forecasts O’Lenic et al (2008) compared the HSS and % non-EC of official (OFF) forecasts made in real-time from 1995 through 2004 with forecasts made using an objective consolidation (CON) of the identical four main forecast tools which were used to prepare the real-time forecasts. Period (1995-2004) mean T, P forecast skills for OFF/CON were 22/26 for T, and 8.8/12.1 for P. CON % non-EC is also higher. CPC began using CON as a first guess in 2006.

21 21 +20% +8% +18% +16% Top 2 rows: 1995-2004 HSS (lines) of 3-month P Outlooks, Official (OFF) and Consolidation (CON). Colors are the fraction of the time non-EC is predicted (%). Bottom row: Difference, CON-OFF (lines and colors). (See O’Lenic et al, 2008) DIF CON HSS DIF CON HSS OFF DIF SPRING SUMMER FALL WINTER FMA, MAM, AMJ MJJ, JJA, JAS ASO, SON, OND NDJ, DJF, JFM DIF

22 22 SPRING SUMMER FALL WINTER FMA, MAM, AMJ MJJ, JJA, JAS ASO, SON, OND NDJ, DJF, JFM +11% +31% +40% DIF CON HSS DIF +55% CON HSS OFF DIF Top 2 rows: 1995-2004 HSS (lines) of 3-month T Outlooks, Official (OFF) and Consolidation (CON). Colors are the fraction of the time non-EC is predicted (%). Bottom row: Difference, CON-OFF (lines and colors). DIF

23 23 GPRA Score Official Skill Metric: 48-Mo. Running Mean U.S. Average T HSS

24 24 What is the Forecast Evaluation Tool (FET)?  A web tool that allows users to interact with a database of CPC 3-Month Mean Temperature, Total Precipitation Outlooks, and verifying observations from1995-present,  A tracker of user preferences.  A self-teaching tool to enable a wide range of users to learn what CPC forecasts are, what they mean, and what their implications are for user applications.  A collector of user requirements.  A community tool for CPC, IRI, IPCC, …???  A laboratory for growing services to users

25 25 http://fet.hwr.arizona.edu/ForecastEvaluationTool/

26 26 Introduction to the FET

27 27 Tutorial

28 28 Look at the latest forecasts

29 29 Look at the latest forecasts

30 30 Interact with observations

31 31

32 32 Las Vegas Historical La Nina Observed P LA NINA NEUTRALEL NINO

33 33 Central Florida Historical El Nino Observed P LA NINA NEUTRAL EL NINO

34 34 Oregon (east) Historical El Nino Observed P PDO - PDO neut PDO +

35 35 Verification(regime, score, years)

36 36 Verification(regime, score, years)

37 37 Frequency of non-EC Forecasts

38 38 Probability of Detection

39 39 False Alarm Rate

40 40 Brier Skill Score

41 41 Ranked Probability Skill Score

42 42 FUTURE: Expand FET’s Capabilities

43 43 CTB User-Centric Forecast Tools Progress Simple Object Access Protocol (SOAP) Tested Secured Go-Ahead to Place FET on NWS Web Operations Center (WOC) Trained CPC Staff in JAVA language Scheduled Ellen Lay Training session in Nov.

44 44 FUTURE of the FET Next 1-4 months: Finalize and implement FET project plan at CPC. Ellen Lay (CLIMAS) to train CPC personnel on FET version control and bug tracking at CPC, November 18-21, 2008. Necessary software (APACHE TOMCAT, JAVA, Desktop View) acquired and installed at CPC. Forecast, observations datasets in-place at CPC. FET code ported to CPC, installed, tested. FET installed to NWS Web Operations Center (WOC) servers

45 45 FUTURE of the FET + In partnership with CLIMAS and community we will add: Other forecasts and organizations Time and space aggregation options Significance tests/cautions to users Requirements requests option Questions option The stakes are high…..

46 46 Source: The Washington Post Outlook Section, July 13, 2008

47 47 Summary Users want partnership, accuracy, specificity, flexibility “Relationship” is synonymous with “partnership” TRUST (honest brokerage) is central to these requirements. Producers must learn WHO users are, HOW they use products and WHAT their evolving requirements are Need to involve users and producers in iteratively optimizing products A continuous flow of requirements from users toward research may avert VOD. Means to fund an ever-expanding, perpetual product suite needed Stakes are high: 6 of top 20 news/media June 2008 sites were weather-related. 10s-100s of B$ at stake. Climate will only add to this.


Download ppt "1 A Forecast Evaluation Tool (FET) for CPC Operational Forecasts Edward O’Lenic Chief, Operations Branch, NOAA-NWS-CPC."

Similar presentations


Ads by Google