CESM Infrastructure Update

Slides:



Advertisements
Similar presentations
CESM Breckenridge Workshop June 20, 2013 CESM Breckenridge Workshop June 20, 2013 Cheryl Craig and Steve Goldhaber with Andrew Gettelman,Julio Bacmeister,
Advertisements

Technology Drivers Traditional HPC application drivers – OS noise, resource monitoring and management, memory footprint – Complexity of resources to be.
Phillip Dickens, Department of Computer Science, University of Maine. In collaboration with Jeremy Logan, Postdoctoral Research Associate, ORNL. Improving.
Multi-Mode Survey Management An Approach to Addressing its Challenges
CCSM Testing Status Tony Craig Lawrence Buja Wei Yu CCSM SEWG Meeting Feb 5, 2003.
Systems Engineering in a System of Systems Context
Integrated Frameworks for Earth and Space Weather Simulation Timothy Killeen and Cecelia DeLuca National Center for Atmospheric Research, Boulder, Colorado.
Mesoscale & Microscale Meteorological Division / NCAR ESMF and the Weather Research and Forecast Model John Michalakes, Thomas Henderson Mesoscale and.
Common Infrastructure for Modeling the Earth
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 18 Slide 1 Software Reuse 2.
WRF-VIC: The Flux Coupling Approach L. Ruby Leung Pacific Northwest National Laboratory BioEarth Project Kickoff Meeting April 11-12, 2011 Pullman, WA.
NSF NCAR | NASA GSFC | DOE LANL ANL | NOAA NCEP GFDL | MIT Adoption and field tests of M.I.T General Circulation Model (MITgcm) with ESMF Chris Hill ESMF.
Metadata Creation with the Earth System Modeling Framework Ryan O’Kuinghttons – NESII/CIRES/NOAA Kathy Saint – NESII/CSG July 22, 2014.
CCSM Software Engineering Coordination Plan Tony Craig SEWG Meeting Feb 14-15, 2002 NCAR.
Quality Attributes of Web Software Applications – Jeff Offutt By Julia Erdman SE 510 October 8, 2003.
An Introduction to Software Architecture
NE II NOAA Environmental Software Infrastructure and Interoperability Program Cecelia DeLuca Sylvia Murphy V. Balaji GO-ESSP August 13, 2009 Germany NE.
ESMF Development Status and Plans ESMF 4 th Community Meeting Cecelia DeLuca July 21, 2005 Climate Data Assimilation Weather.
CESM/RACM/RASM Update May 15, Since Nov, 2011 ccsm4_0_racm28:racm29:racm30 – vic parallelization – vic netcdf files – vic coupling mods and “273.15”
Service-enabling Legacy Applications for the GENIE Project Sofia Panagiotidi, Jeremy Cohen, John Darlington, Marko Krznarić and Eleftheria Katsiri.
DOE BER Climate Modeling PI Meeting, Potomac, Maryland, May 12-14, 2014 Funding for this study was provided by the US Department of Energy, BER Program.
Computational Design of the CCSM Next Generation Coupler Tom Bettge Tony Craig Brian Kauffman National Center for Atmospheric Research Boulder, Colorado.
Initial Results from the Integration of Earth and Space Frameworks Cecelia DeLuca/NCAR, Alan Sussman/University of Maryland, Gabor Toth/University of Michigan.
Overview of ESMF in the Community Climate System Model (CCSM) Erik Kluzek NCAR -- CCSM Software Engineering Group (CSEG) Erik Kluzek NCAR -- CCSM Software.
ESMF Application Status GMAO Seasonal Forecast NCAR/LANL CCSM NCEP Forecast GFDL FMS Suite MITgcm NCEP/GMAO Analysis Climate Data Assimilation.
Overarching System and NEMS Cecelia DeLuca NOAA Environmental Software Infrastructure and Interoperability (NESII) NOAA ESRL/University of Colorado NGGPS.
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
Lessons learned from building and managing the Community Climate System Model David Bailey PCWG liaison (NCAR) Marika Holland PCWG co-chair (NCAR) Elizabeth.
CSEG Update Mariana Vertenstein CCSM Software Engineering Group Mariana Vertenstein CCSM Software Engineering Group.
ESPC Infrastructure Update on HYCOM in CESM (Alex Bozec, Fei Liu, Kathy Saint, Mat Rothstein, Mariana Vertenstein, Jim Edwards, …) Update on ESMF integration.
CESM/ESMF Progress Report Mariana Vertenstein NCAR Earth System Laboratory CESM Software Engineering Group (CSEG) NCAR is sponsored by the National Science.
DTC HWRF Task AOP2009 & AOP /01/2009 Ligia Bernardet.
1 Geospatial and Business Intelligence Jean-Sébastien Turcotte Executive VP San Francisco - April 2007 Streamlining web mapping applications.
DataNet – Flexible Metadata Overlay over File Resources Daniel Harężlak 1, Marek Kasztelnik 1, Maciej Pawlik 1, Bartosz Wilk 1, Marian Bubak 1,2 1 ACC.
Earth System Modeling Framework Status Cecelia DeLuca NOAA Cooperative Institute for Research in Environmental Sciences University of Colorado, Boulder.
Regional Models in CCSM CCSM/POP/ROMS: Regional Nesting and Coupling Jon Wolfe (CSEG) Mariana Vertenstein (CSEG) Don Stark (ESMF)
Components, Coupling and Concurrency in the Earth System Modeling Framework N. Collins/NCAR, C. DeLuca/NCAR, V. Balaji/GFDL, G. Theurich/SGI, A. da Silva/GSFC,
Pg 1 A Loosely Coupled Ocean-Atmosphere Ensemble Assimilation System. Tim Hoar, Nancy Collins, Kevin Raeder, Jeffrey Anderson, NCAR Institute for Math.
A Portable Regional Weather and Climate Downscaling System Using GEOS-5, LIS-6, WRF, and the NASA Workflow Tool Eric M. Kemp 1,2 and W. M. Putman 1, J.
Towards development of a Regional Arctic Climate System Model --- Coupling WRF with the Variable Infiltration Capacity land model via a flux coupler Chunmei.
Earth System Modeling Framework Python Interface (ESMP) October 2011 Ryan O’Kuinghttons Robert Oehmke Cecelia DeLuca.
CCSM Portability and Performance, Software Engineering Challenges, and Future Targets Tony Craig National Center for Atmospheric Research Boulder, Colorado,
CCA Common Component Architecture CCA Forum Tutorial Working Group CCA Status and Plans.
ESMF Strategic Discussion Cecelia DeLuca NOAA ESRL/University of Colorado ESMF Executive Board/Interagency Meeting June 12, 2014.
CCSM Performance, Successes and Challenges Tony Craig NCAR RIST Meeting March 12-14, 2002 Boulder, Colorado, USA.
On the Road to a Sequential CCSM Robert Jacob, Argonne National Laboratory Including work by: Mariana Vertenstein (NCAR), Ray Loy (ANL), Tony Craig (NCAR)
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
Running CESM An overview
Process Asad Ur Rehman Chief Technology Officer Feditec Enterprise.
Enhancing Security in Enterprise Distributed Real-time and Embedded Systems using Domain-specific Modeling Akshay Dabholkar, Joe Hoffert, Aniruddha Gokale,
CCSM Software Engineering Update Tony Craig CCSM SEWG Meeting Feb 4, 2003.
Emergence of a Common Modeling Architecture for Earth System Science American Geophysical Union December 13, 2010 Cecelia DeLuca NOAA/CIRES.
State of ESMF: The NUOPC Layer Gerhard Theurich NRL/SAIC ESMF Executive Board / Interagency Working Group Meeting June 12, 2014.
NOAA Environmental Modeling System Cecelia DeLuca NOAA Environmental Software Infrastructure and Interoperability (NESII) NOAA ESRL/University of Colorado.
A TIME-GCM CAM Multi-executable Coupled Model Using ESMF and InterComm Robert Oehmke, Michael Wiltberger, Alan Sussman, Wenbin Wang, and Norman Lo.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech 5 th GO-ESSP Community Meeting.
Coupled HYCOM in CESM and ESPC Alexandra Bozec, Eric P. Chassignet.
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca Dr. Robert Detrick visit March 28, 2012
Overview of the CCSM CCSM Software Engineering Group June
Experiences and Decisions in Met Office coupled ESM Development
Towards development of a Regional Arctic Climate System Model ---
ESPC Air-Ocean-Land-Ice Global Coupled Prediction
ESMF Governance Cecelia DeLuca NOAA CIRES / NESII April 7, 2017
Software Practices for a Performance Portable Climate System Model
Performance of the VIC land surface model in coupled simulations
Mariana Vertenstein (CGD)
Joint GEOS-Chem and NCAR Modeling Workshop:
Mariana Vertenstein CCSM Software Engineering Group NCAR
Progress of Interactions Among CCSM and Other Modeling Efforts
A brief introduction to NEMS
Presentation transcript:

CESM Infrastructure Update Mariana Vertenstein mvertens@ucar.edu CESM Software Engineering Group National Center For Atmospheric Research CESM is primarily sponsored by the National Science Foundation and the Department of Energy

Outline New approach to infrastructure development Common Infrastructure for Modeling the Earth - CIME New coupling complexity New components, routing complexity grids Challenges of data assimilation ESMF collaboration NUPOC, On-line regridding New Infrastructure Capabilities (1) Statistical Ensemble Test (2) Creation of parallel workflow capabilities (3) PIO2

Common Infrastructure for Modeling the Earth CIME A New Approach for Earth System Modeling – the CESM example https://github.com/ESMCI/cime

In past infrastructure (no IP) tied to science development (has IP) Ocean (POP, Data) SEA ICE (CICE, Data) COUPLER Atmosphere (CAM, Data) Land (CLM, DATA) Wave (WW3, DATA) River (RTM, DATA) Land Ice (CISM)

Why CIME? Facilitate infrastructure modernization as a collaborative project (e.g. CESM infrastructure) Response to February summit of US Global Change Research Program (USGCRP) / Interagency Group on Integrative Modeling (IGIM) as a positive outcome from the February Summit IGIM is charged with coordinating global change-related modeling activities across the Federal Government and providing guidance to USGCRP on modeling priorities. Enable separation of infrastructure (no intellectual property) versus scientific development codes (intellectual property must be protected) Eliminate duplication of efforts

CIME current steps forward….. ALL CESM infrastructure to PUBLIC github repository This will facilitate AND encourage outside collaboration frequent feedback on infrastructure development quick problem resolution rapid improvement in the productivity, reliability and extensibility of the CIME infrastructure CIME can developed and tested as a stand-alone system – independent of prognostic components

Old paradigm – everything in restricted developer repository Infrastructure Restricted Subversion Repository - All model components Restricted Subversion Repository Driver-Coupler Code Share Code Scripts System and Unit Testing Mapping Utilities CAM (prognostic) DATM (data) SATM (stub) XATM (cpl test) ATM Models

New paradigm – all infrastructure is Open Source IP still in place for prognostic components PUBLIC Open Source Github Repository - Only prognostic components Restricted Subversion Repository Driver-Coupler Share Code Scripts Mapping Utilities System/Unit Testing All Data Models All Stub Models All cpl-test Models CAM CLM CICE POP (MPAS) RTM(MOSART) CISM Prognostic

PUBLIC Open Source Github Repository - CIME Infrastructure can be used to facilitate releases and external collaborations Infrastructure PUBLIC Open Source Github Repository - Driver-Coupler (ESMF Collaboration) Share Code Scripts System/Unit testing Mapping Utilities All Data Models All Stub Models All cpl-test Models Prognostic Components e.g. CESM ESMF/NUOPC HYCOM

CIME implementation Stand-alone capability New unit testing framework CIME run and tested “stand-alone” with either all data models or all stub-models or all “test-cpl” models New unit testing framework new unit tests in coupler Framework can also be applied to prognostic components Consolidation of separate externals Each current part of CIME was previously developed independently and often led to inconsistencies Now CIME is a SINGLE entity and ensures consistency among its various parts This simplifies and adds robustness to development process

Coupling Challenges New component, routing and grid complexity, DART data assimilation

Coupling Complexity Currently have 7 components Ocean (POP, Data) SEA ICE (CICE, Data) COUPLER Atmosphere (CAM, Data) Land (CLM, DATA) Wave (WW3, DATA) River (RTM, DATA) Land Ice (CISM)

Routing and Regridding Complexity Each component can run on its own grid – only assumption is that ocn/ice are on the same grid Multiple grids supported Regular lat/lon Dipole, Tripole (ocn-pop, ice-cice) Hexagonal (regular and unstructured Voronoi meshes) MPAS grids (atm dycore, ocn, ice, land-ice) Coupler is responsible for regridding currently using mapping files that are generated offline with ESMF parallel regridding tool Fluxes mapped conservatively States mapped with either bi-linear or higher order non-conservative Components communicate with coupler at potentially different frequencies and with unique routine patterns

Coupling Complexity (Routing) Ocean (POP, Data) SEA ICE (CICE, Data) COUPLER Atmosphere (CAM, Data) Land Ice (CISM)

Coupling Complexity (Routing) Ocean (POP, Data) SEA ICE (CICE, Data) COUPLER Land (CLM, DATA) River (RTM, DATA)

Coupling Complexity (Routing) COUPLER Atmosphere (CAM, Data) Land (CLM, DATA) River (RTM, DATA) Land Ice (CISM)

Multiple instance capability DART data assimilation ocn obs atm obs Atmosphere (CAM) Land (CLM) Ocean (POP) SEA ICE (CICE) COUPLER River-Runoff DART

Next steps in coupling with data assimilation All data assimilation currently done via files and DART and CESM are separate executables CESM must be stopped and restarted every data assimilation interval (6 hours) – extremely inefficient and expensive use of system resources Limitations of only 1 coupler, but multiple component instances New project enable DART to be a component of CESM coupled system each CESM component will have pause/resume capability and will be able to start up from a restart file during the model run Extend coupler capability to permit multiple couplers within single executable

ESMF Collaboration NUOPC and Online Regridding

Two ESMF/CESM Collaborations Online regridding ESMF is collaborating with CSEG and this will be brought into CIME Introduction of “ESMF/NUOPC” capability in CESM/CIME Vertenstein is co-PI on ESPC proposal “An Integration and Evaluation Framework for ESPC Coupled Models” Import the standalone version of HYCOM into the validated NUOPC version of the CESM coupled system and test using the CORE forcing Run reference CESM configurations (500 years present day climate + IPCC scenarios) for both POP and HYCOM

NUOPC and CESM What is NUOPC? Goal for Driver Goal for Components: NUOPC layer “generic components” are templates that encode rules for drivers, models, mediators (custom coupling code) and connectors (for data transfer) Goal for Driver Maintain a single CESM driver – but restructure it to accommodate both MCT and NUOPC component interfaces Goal for Components: Implement ESMF-based NUOPC components as a CESM option – but build on existing ESMF component interfaces

Redesign of cpl7 as first step Why? Driver code was one large routine (6K loc)and hard-coded to contain MCT data types Difficult to understand, modify and add an alternative coupling architecture What was the redesign? Introduced a new abstraction layer between driver and components – driver has no reference to MCT or ESMF types Much easier to incorporate new ESMF/NUOPC driver components Permits backwards compatibility and memory sharing between MCT and ESMF data structures

MCT Hub<-> CAM Exchange ESMF Hub<-> CAM Exchange Original Redesigned CAM DRIVER Component-type based MCT Hub<-> CAM Exchange or ESMF Hub<-> CAM Exchange Component_type <-> ESMF Component_type <-> MCT CAM MCT  CAM DRIVER MCT based MCT Hub<-> CAM Exchange MCT  ESMF MCT  CAM ESMF  CAM ESMF  CAM

Current and Future Work Current Status: NUOPC implementation complete for all CESM components AND HYCOM Modified the data exchange (between coupler/Mediator and components) to use NUOPC Fields Future Work: Clean up and prepare NUOPC version for wider use Merge code back to CIME and reconcile with other developmental changes as well as MCT implementation Performance evaluation

ESMF online regridding As more regionally refined grids are introduced (e.g. MPAS, SE) – need to minimize the number of mapping files that are needed simplify and streamline workflow for generating new user grid configurations Will be a requirement for run-time adaptive mesh refinement ESMF is the only tool that currently delivers this capability Status: Prototype implementation has been done and is being updated for newest coupler

New Infrastructure Capabilities (1) Statistical Ensemble Test

CESM Ensemble Consistency Test Motivation: Ensure that changes during the CESM development cycle (code modifications, compiler changes, new machine architectures) do not adversely effect the code Question: Is the new data statistically significant from the old one Old Method: compare multiple long simulations – time consuming and subjective New Method: evaluate new data in the context of an ensemble of CESM runs

CESM Ensemble Consistency Test Part 1: Create “truth” ensemble of 1-year CESM runs (151) Use “accepted” machine and “accepted” software stack Differ by O (10-14) perturbations in initial atmospheric temp. Part 2: Create “accepted” statistical distribution Statistics based on ensemble Summary file included with CESM release Part 3: Evaluate “new” runs (new platform, code base, …) Create 3 “new” runs (randomly selected i.c. from ensemble) Principal Component Analysis (PCA)-based testing Provides false positive rates Part 1: Done for CESM release 1-deg atmosphere model (F-case): 120 variables Looking at annual averages Part 2: statistics summary file Included in CESM 1.3.x release – parallel python code ~20 min Part 3: Python tool included in release that a user can run

CESM Ensemble Consistency Test Many uses: Port-verification (new CESM-supported architectures, heterogeneous computing platforms) Sanity check on climate similarity for new CESM sytem snapshots (caught recent bug that would not have been caught before!!!!) Exploration of new algorithms, solvers, compiler options, … Evaluation of data compression on CESM data Heterogeneous computing (GPU/CPU)

New Infrastructure Capabilities (2) New End-to-End workflow Parallel Post-Processing as part of the model run

New Infrastructure Capabilities (3) New Parallel IO Libarary (PIO2)

New Parallel IO Library : PIO2 PIO2 rewrite has new C language API added (F90 API retained) decomposition option improves scalability data aggregation which improves performance new testing framework Provides higher performance through data aggregation new subset rearranger provides higher scalability Provides more options for IO performance tuning but also provides tools to make tuning easier

Subset rearranger gives better scaling Data rearranged to 32 IO tasks Box rearranger gives optimal data layout Example decomposition: CLM data on 2048 tasks.

Acknowledgements (plus many more) ESMF Collaboration Cecelia Deluca, Fei Liu, Gerhard Theurich, Peggy Li, Robert Oehmke, Mathew Rothstein, Tony Craig, Jim Edwards Statistical Ensemble Test Allison Baker, Dorrit Hammerling, Mike Levy, Doug Nychka, John Dennis, Joe Tribbia, Dave Williamson, Jim Edwards Parallel IO Jim Edwards, John Dennis, Jayesh Krishna Data Assimilation Alicia Karspeck, Nancy Collins, Tim Hoar, Kevin Reader, Jeff Anderson, Tony Craig Parallel Workflow Alice Bertini, Sheri Mickelson, Jay Schollenberger CSEG – Ben Andre, David Bailey, Alice Bertini, Cheryl Craig, Tony Craig, Brian Eaton, Jim Edwards, Erik Kluzek, Mike Levy, Bill Sacks, Sean Santos, Jay Schollenberger