Data Requirements for Climate and Carbon Research

Slides:



Advertisements
Similar presentations
NCAR The Earth System Grid (ESG) & The Community Data Portal (CDP) (NCARs Data & GriD Efforts) for COMMISSION FOR BASIC SYSTEMS INFORMATION SYSTEMS and.
Advertisements

NCAR Cyberinfrastructure for Earth System Modeling Don Middleton NCAR Scientific Computing Division APAN eScience Workshop, Honolulu January 28, 2004.
NG-CHC Northern Gulf Coastal Hazards Collaboratory Simulation Experiment Integration Sandra Harper 1, Manil Maskey 1, Sara Graves 1, Sabin Basyal 1, Jian.
A. Sim, CRD, L B N L 1 ANI and Magellan Launch, Nov. 18, 2009 Climate 100: Scaling the Earth System Grid to 100Gbps Networks Alex Sim, CRD, LBNL Dean N.
DOE Global Modeling Strategic Goals Anjuli Bamzai Program Manager Climate Change Prediction Program DOE/OBER/Climate Change Res Div
Earth System Curator Spanning the Gap Between Models and Datasets.
Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Global Climate Modeling Research John Drake Computational Climate Dynamics Group Computer.
Toni Saarinen, Tite4 Tomi Ruuska, Tite4 Earth System Grid - ESG.
The Earth System Grid Discovery and Semantic Web Technologies Line Pouchard Oak Ridge National Laboratory Luca Cinquini, Gary Strand National Center for.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
TPAC Digital Library Talk Overview Presenter:Glenn Hyland Tasmanian Partnership for Advanced Computing & Australian Antarctic Division Outline: TPAC Overview.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
CCSM Portal/ESG/ESGC Integration (a PY5 GIG project) Lan Zhao, Carol X. Song Rosen Center for Advanced Computing Purdue University With contributions by:
NCAR NCAR Data and Grid Efforts: The Earth System Grid & The Community Data Portal Don Middleton NCAR Scientific Computing Division CAS2003 September 11,
Presented by The Earth System Grid: Turning Climate Datasets into Community Resources David E. Bernholdt, ORNL on behalf of the Earth System Grid team.
Ian Foster Argonne National Lab University of Chicago Globus Project The Grid and Meteorology Meteorology and HPN Workshop, APAN.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
ESG The Earth System Grid (ESG) Presented by Don Middleton & Luca Cinquini NCAR Scientific Computing Division On Behalf of the ESG Team SCD Executive Committee.
Mathematics and Computer Science & Environmental Research Divisions ARGONNE NATIONAL LABORATORY Regional Climate Simulation Analysis & Vizualization John.
The Earth System Grid (ESG) Goals, Objectives and Strategies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
Integrated Model Data Management S.Hankin ESMF July ‘04 Integrated data management in the ESMF (ESME) Steve Hankin (NOAA/PMEL & IOOS/DMAC) ESMF Team meeting.
Intergrid KoM Santander 22 june, 2006 E-Infraestructure shared between Europe and Latin America José Manuel Gutiérrez
- EGU 2010 ESSI May Building on the CMIP5 effort to prepare next steps : integrate community related effort in the every day workflow to.
Soil and Water Conservation Modeling: MODELING SUMMIT SUMMARY COMMENTS Dennis Ojima Natural Resource Ecology Laboratory COLORADO STATE UNIVERSITY 31 MARCH.
The Earth System Grid: A Visualisation Solution Gary Strand.
Web Portal Design Workshop, Boulder (CO), Jan 2003 Luca Cinquini (NCAR, ESG) The ESG and NCAR Web Portals Luca Cinquini NCAR, ESG Outline: 1.ESG Data Services.
The Earth System Grid (ESG) Computer Science and Technologies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
NIEeS Workshop, Cambridge (UK), Sep 2002 Luca Cinquini for the Earth System Grid METADATA DEVELOPMENT for the EARTH SYSTEM GRID Luca Cinquini (SCD/NCAR)
The Digital Library for Earth System Science: Contributing resources and collections GCCS Internship Orientation Holly Devaul 19 June 2003.
Metadata Standards for Gridded Climate Data in the Earth System Grid Robert Drach LLNL/PCMDI UCRL-PRES
- Vendredi 27 mars PRODIGUER un nœud de distribution des données CMIP5 GIEC/IPCC Sébastien Denvil Pôle de Modélisation, IPSL.
GEON2 and OpenEarth Framework (OEF) Bradley Wallet School of Geology and Geophysics, University of Oklahoma
ESG Observational Data Integration Presented by Feiyi Wang Technology Integration Group National Center of Computational Sciences.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
Access Control for NCAR Data Portals A report on work in progress about the future of the NCAR Community Data Portal Luca Cinquini GO-ESSP Workshop, 6-8.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
1 Overall Architectural Design of the Earth System Grid.
1 Earth System Grid Center For Enabling Technologies (ESG-CET) Introduction and Overview Dean N. Williams, Don E. Middleton, Ian T. Foster, and David E.
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
1 Summary. 2 ESG-CET Purpose and Objectives Purpose  Provide climate researchers worldwide with access to data, information, models, analysis tools,
Fire Emissions Network Sept. 4, 2002 A white paper for the development of a NSF Digital Government Program proposal Stefan Falke Washington University.
CUAHSI HIS: Science Challenges Linking small integrated research sites (
2. WP9 – Earth Observation Applications ESA DataGrid Review Frascati, 10 June Welcome and introduction (15m) 2.WP9 – Earth Observation Applications.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
SCD User Briefing The Community Data Portal and the Earth System Grid Don Middleton with presentation material developed by Luca Cinquini, Mary Haley,
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Data Requirements for Climate and Carbon Research John Drake, Climate Dynamics Group Computer.
Climate-SDM (1) Climate analysis use case –Described by: Marcia Branstetter Use case description –Data obtained from ESG –Using a sequence steps in analysis,
The Global Organization for Earth System Science Portals Don Middleton National Center for Atmospheric Research; Boulder, Colorado, USA Scientific Computing.
The NOAA Operational Model Archive and Distribution System NOMADS CEOS-Grid Application Status Report Glenn K. Rutledge NOAA NCDC CEOS WGISS-19 Cordoba,
Metadata Support for Model Intercomparison Projects Sylvia Murphy: Cecelia DeLuca: Julien.
DataGrid France 12 Feb – WP9 – n° 1 WP9 Earth Observation Applications.
Clouds , Grids and Clusters
The Earth System Grid: A Visualisation Solution
DOE Facilities - Drivers for Science: Experimental and Simulation Data
Collaborations and Interactions with other Projects
improve the efficiency, collaborative potential, and
Grid Portal Services IeSE (the Integrated e-Science Environment)
C2CAMP (A Working Title)
Software Practices for a Performance Portable Climate System Model
Grid Computing B.Ramamurthy 9/22/2018 B.Ramamurthy.
Multidimensional Data and GIS
Prepared by: Jennifer Saleem Arrigo, Program Manager
Geospatial and Problem Specific Semantics Danielle Forsyth, CEO and Co-Founder Thetus Corporation 20 June, 2006.
HAO/SCD: VO, metadata, catalogs, ontologies, querying
About Thetus Thetus develops knowledge discovery and modeling infrastructure software for customers who: Have high value data that does not neatly fit.
Metadata Development in the Earth System Curator
Data Management Components for a Research Data Archive
Presentation transcript:

Data Requirements for Climate and Carbon Research David C. Bader Chief Scientist, DOE Climate Change Prediction Program Presentation Material Courtesy of: John Drake, ORNL Don Middleton, National Center for Atmospheric Research and CCPP Scientists UCRL-PRES-202932

Why is DOE Interested in Climate?

Courtesy Warren Washington

IPCC Fourth Assessment Runs Plan: Commitment Runs : B1 Scenario simulated with ensemble calculations Dedicated 6 (32way) nodes of Cheetah for Sept-Nov 2003 Dedicated 12 (32way) nodes of Cheetah for Mar-July 2004 Cray X1 Science runs in 2004

Baseline Numbers T42 CCSM (current, 280km) T85 CCSM (140km) 7.5GB/yr, 100 years -> .75TB T85 CCSM (140km) 29GB/yr, 100 years -> 2.9TB T170 CCSM (70km) 110GB/yr, 100 years -> 11TB

Diagonstic Analysis of Coupled Models Courtesy PCMDI Courtesy CCSM/AMWG

Tools

The Earth System Grid http://www.earthsystemgrid.org U.S. DOE SciDAC funded R&D effort Build an “Earth System Grid” that enables management, discovery, distributed access, processing, & analysis of distributed terascale climate research data A “Collaboratory Pilot Project” Build upon ESG-I, Globus Toolkit, DataGrid technologies, and deploy Potential broad application to other areas

ESG Team LLNL/PCMDI USC/ISI ANL NCAR LBNL ORNL Bob Drach Dean Williams (PI) USC/ISI Anne Chervenak Carl Kesselman (Laura Perlman) NCAR David Brown Luca Cinquini Peter Fox Jose Garcia Don Middleton (PI) Gary Strand ANL Ian Foster (PI) Veronika Nefedova (John Bresenhan) (Bill Allcock) LBNL Arie Shoshani Alex Sim ORNL David Bernholdte Kasidit Chanchio Line Pouchard

Capacity-related Improvements Increased turnaround, model development, ensemble of runs Increase by a factor of 10, linear data Current T42 CCSM 7.5GB/yr, 100 years -> .75TB * 10 = 7.5TB Many more runs, need to keep track of them

Capability-related Improvements Spatial Resolution: T42 -> T85 -> T170 Increase by factor of ~ 10-20, linear data Temporal Resolution: Study diurnal cycle, 3 hour data Increase by factor of ~ 4, linear data Note potential trend toward process-oriented analysis CCM3 at T170 (70km)

Capability-related Improvements Quality: Improved boundary layer, clouds, convection, ocean physics, land model, river runoff, sea ice Increase by another factor of 2-3, data flat Scope: Atmospheric chemistry (sulfates, ozone…), biogeochemistry (carbon cycle, ecosystem dynamics), middle Atmosphere Model… Increase by another factor of 10+, linear data

Model Improvements cont. Grand Total: Increase compute by a Factor O(1000-10000) Trends Increased space and time resolution Increased complexity Increased scope (e.g. biogeochemical) Capability advances will lead to analysis that becomes increasingly process-oriented. Analysis may become more exploratory and “intense”. Many more runs and we must keep track of them.

ESG: Challenges Enabling the simulation and data management team Enabling the core research community in analyzing and visualizing results Enabling broad multidisciplinary communities to access simulation results And as we’re able to compute more and quicker, the workflow among these groups becomes ever more important John Shalf and others made some good points in this area in yesterday’s vis session Note observational data We need integrated scientific work environments that enable smooth WORKFLOW for knowledge development: computation, collaboration & collaboratories, data management, access, distribution, analysis, and visualization.

ESG: Strategies Move data a minimal amount, keep it close to computational point of origin when possible Data access protocols, distributed analysis When we must move data, do it fast and with a minimum amount of human intervention Storage Resource Management, fast networks Keep track of what we have, particularly what’s on deep storage Metadata and Replica Catalogs Harness a federation of sites, web portals Globus Toolkit -> The Earth System Grid -> The UltraDataGrid Coupled hardware resources for data Distributed data access protocols It’s already very hard to keep track of all the different runs, for different scenarios, at different sites. Metadata will enable tracking as well as enhanced analysis capabilities Even with an ESS, we will have multiple sites

HRM aka “DataMover” Running well across DOE/HPSS systems New component built that abstracts NCAR Mass Storage System Defining next generation of requirements with climate production group First “real” usage

OPeNDAP An Open Source Project for a Network Data Access Protocol (originally DODS, the Distributed Oceanographic Data System)

ESG: Metadata Services EXTRACTION DISPLAY BROWSING QUERY ESG CLIENTS API & USER INTERFACES Data & Metadata Catalog Dublin Core Database COARDS mirror XML Files COMMENTS METADATA HOLDINGS ANNOTATION VALIDATION METADATA ACCESS (update, insert, delete, query) SERVICE TRANSLATION LIBRARY CORE METADATA SERVICES AGGREGATION DISCOVERY METADATA & DATA REGISTRATION PUBLISHING HIGH LEVEL METADATA SERVICES SEARCH & DISCOVERY ADMINISTRATION BROWSING & DISPLAY ANALYSIS & VISUALIZATION

Collaborations & Relationships CCSM Data Management Group The Globus Project Other SciDAC Projects: Climate, Security & Policy for Group Collaboration, Scientific Data Management ISIC, & High-performance DataGrid Toolkit OPeNDAP/DODS (multi-agency) NSF National Science Digital Libraries Program (UCAR & Unidata THREDDS Project) U.K. e-Science and British Atmospheric Data Center NOAA NOMADS and CEOS-grid Earth Science Portal group (multi-agency, intnl.)