- Vendredi 27 mars 2009 1 PRODIGUER un nœud de distribution des données CMIP5 GIEC/IPCC Sébastien Denvil Pôle de Modélisation, IPSL.

Slides:



Advertisements
Similar presentations
Earth System Curator Spanning the Gap Between Models and Datasets.
Advertisements

Metadata Development in the Earth System Curator Spanning the Gap Between Models and Datasets Rocky Dunlap, Georgia Tech.
Preservation and Long Term Access of Data at the World Data Centre for Climate Frank Toussaint N.P. Drakenberg, H. Höck, M. Lautenschlager, H. Luthardt,
High Performance Computing Course Notes Grid Computing.
ESCI/CMIP5 Tools - Jeudi 2 octobre CMIP5 Tools Earth System Grid-NetCDF4- CMOR2.0-Gridspec-Hyrax …
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Application of GRID technologies for satellite data analysis Stepan G. Antushev, Andrey V. Golik and Vitaly K. Fischenko 2007.
UMIACS PAWN, LPE, and GRASP data grids Mike Smorul.
The Earth System Grid Discovery and Semantic Web Technologies Line Pouchard Oak Ridge National Laboratory Luca Cinquini, Gary Strand National Center for.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
CCSM Portal/ESG/ESGC Integration (a PY5 GIG project) Lan Zhao, Carol X. Song Rosen Center for Advanced Computing Purdue University With contributions by:
1 Eric Guilyardi and the Metafor team Common Metadata for Climate Modelling Digital Repositories Metafor Dissemination Workshop Abingdon, 14 March 2011.
IS-ENES [ees-enes] InfraStructure for the European Network for Earth System Modelling IS-ENES will develop a virtual Earth System Modelling Resource Centre.
CIM – The Common Information Model in Climate Research
Presented by The Earth System Grid: Turning Climate Datasets into Community Resources David E. Bernholdt, ORNL on behalf of the Earth System Grid team.
GESIS Dr. Maximilian Stempfhuber Head of Research and Development Social Science Information Centre, Bonn, Germany How to deal with heterogeneity when.
Ian Foster Argonne National Lab University of Chicago Globus Project The Grid and Meteorology Meteorology and HPN Workshop, APAN.
Initiatives toward Climate Services in France and in the European Communities C. Déandreis (CNRS/IPSL); M. Plieger and W. Som de Cerff (KNMI); Ph. Dandin,
Planning for Arctic GIS and Geographic Information Infrastructure Sponsored by the Arctic Research Support and Logistics Program 30 October 2003 Seattle,
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Mathematics and Computer Science & Environmental Research Divisions ARGONNE NATIONAL LABORATORY Regional Climate Simulation Analysis & Vizualization John.
The Earth System Grid (ESG) Goals, Objectives and Strategies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
Michael Lautenschlager World Data Center Climate Model and Data / Max-Planck-Institute for Meteorology German Climate Computing Centre (DKRZ)
Research and Educational Networking and Cyberinfrastructure Russ Hobby, Internet2 Dan Updegrove, NLR University of Kentucky CI Days 22 February 2010.
- IS-ENES SA2/JRA4 24 th February IS-ENES SA2/JRA4 IPSL Status/Questions Sébastien Denvil, Ashish Bhardwaj, Mark Morgan IPSL, Global Climate Modeling.
기후모델 : 기후변화연구의 인프라 Climate Model: Infrastructure for Climate Change Research Wonsun Park Leibniz Institute of Marine Sciences Kiel, Germany KOFST Ultra.
ExArch: Climate analytics on distributed exascale data archives Martin Juckes, V. Balaji, B.N. Lawrence, M. Lautenschlager, S. Denvil, G. Aloisio, P. Kushner,
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks G. Quigley, B. Coghlan, J. Ryan (TCD). A.
Data Publication and Quality Control Procedure for CMIP5 / IPCC-AR5 Data WDC Climate / DKRZ:
A Flexible Component based Access Control Architecture for OPeNDAP Services Philip Kershaw STFC Rutherford Appleton Laboratory.
Federated Discovery and Access in Astronomy Robert Hanisch (NIST), Ray Plante (NCSA)
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
The Climate-G testbed towards a large scale data sharing environment for climate change S. Fiore Scientific Computing and Operations Division, CMCC, Italy.
- EGU 2010 ESSI May Building on the CMIP5 effort to prepare next steps : integrate community related effort in the every day workflow to.
Climate Change Working Group (CCWG) July, 2004 Co-chairs: Gerald A. Meehl, Ben Santer, and Warren Washington.
1 Earth System Modeling Framework Documenting and comparing models using Earth System Curator Sylvia Murphy: Julien Chastang:
The Earth System Grid (ESG) Computer Science and Technologies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
The Climate-G testbed towards a large scale data sharing environment for climate change S. Fiore Scientific Computing and Operations Division, CMCC, Italy.
IPCC TGICA and IPCC DDC for AR5 Data GO-ESSP Meeting, Seattle, Michael Lautenschlager World Data Center Climate Model and Data / Max-Planck-Institute.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
WP6/SA2: Access to IS-ENES Data Federation SA2 is a European distributed data infrastructure providing access to data from ESM simulations produced in.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
1 Overall Architectural Design of the Earth System Grid.
1 Gateways. 2 The Role of Gateways  Generally associated with primary sites in ESG-CET  Provides a community-facing web presence  Can be branded as.
1 Earth System Grid Center For Enabling Technologies (ESG-CET) Introduction and Overview Dean N. Williams, Don E. Middleton, Ian T. Foster, and David E.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Earth System Curator and Model Metadata Discovery and Display for CMIP5 Sylvia Murphy and Cecelia Deluca (NOAA/CIRES) Hannah Wilcox (NCAR/CISL) Metafor.
1 Summary. 2 ESG-CET Purpose and Objectives Purpose  Provide climate researchers worldwide with access to data, information, models, analysis tools,
1 Earth System Grid Center for Enabling Technologies (ESG-CET) Overview ESG-CET Team Climate change is not only a scientific challenge of the first order.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
IPCC WG II + III Requirements for AR5 Data Management GO-ESSP Meeting, Paris, Michael Lautenschlager, Hans Luthardt World Data Center Climate.
Application of RDF-OWL in the ESG Ontology Sylvia Murphy: Julien Chastang: Luca Cinquini:
Jost von Hardenberg ISAC-CNR, Torino, Italy with Paolo Davini, Susanna Corti, and many others EUDAT User Forum, Rome,Italy 3-4 February, 2016.
What was done for AR4. Software developed for ESG was modified for CMIP3 (IPCC AR4) Prerelease ESG version 1.0 Modified data search Advance search Pydap.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Metadata Support for Model Intercomparison Projects Sylvia Murphy: Cecelia DeLuca: Julien.
DataGrid France 12 Feb – WP9 – n° 1 WP9 Earth Observation Applications.
AP7/AP8: Long-Term Archival of CMIP6 Data
World Conference on Climate Change October 24-26, 2016 Valencia, Spain
Data Citation Service for CMIP6 and IPCC DDC Aspects
Data Requirements for Climate and Carbon Research
Gridifying the LHCb Monte Carlo production system
Metadata Development in the Earth System Curator
Data Management Components for a Research Data Archive
Presentation transcript:

- Vendredi 27 mars PRODIGUER un nœud de distribution des données CMIP5 GIEC/IPCC Sébastien Denvil Pôle de Modélisation, IPSL

- Vendredi 27 mars Context : countdown of the GIEC/IPCC report  End of 2009  Fall 2010 : Climate simulations  End of 2010  ? : Data Distribution  End of 2010  Early 2012 : Scientific publications  Early 2013 : Report publication GIECC/IPCC AR5 (Assessment Report #5)  Octobre 2013 : Nobel price

- Vendredi 27 mars Context : National and European Project PRODIGUER : project submitted in september 2008 to the GIS climat In the wake of IS-ENES (FP7), Virtual Earth System Modeling resources Centre, Metadata standard and Metafor (FP7) metadata standard for climate modeling Implementation of these tools at national level and integration to International effort Must be done in close collaboration with national computing centers

- Vendredi 27 mars ESG/CMIP5 Timeline 2008: Design and implement core functionality:  Browse and search  Registration  Single sign-on / security  Publication  Distributed metadata  Server-side processing Early 2009: Testbed  By early 2009 it is expected to include at least seven centres in the US, Europe and Japan:  Program for Climate Model Diagnosis and Intercomparison - PCMDI (U.S.),  National Centre for Atmospheric Research - NCAR (U.S.),  Geophysical Fluid Dynamics Laboratory - GFDL (U.S.),  Oak Ridge National Laboratory - ORNL (U.S.),  British Atmosphere Data Centre - BADC (U.K.),  Max Planck Institute for Meteorology - MPI (Germany),  The University of Tokyo Centre for Climate System Research (Japan). 2009: Deal with system integration issues and develop production system.  By summer 2009, the hardware and software requirements will be provided to centres that want to be Nodes. 2010: Modelling centres publish data : Research and journal articles submissions 2013: IPCC Report

- Vendredi 27 mars AR5 open issues What are the set of runs to be done and, derived from that, the expected data volumes we can expect? Expected participants – where will data be hosted? (Who is going to step up and host the data nodes, and provide the level of support expect in terms of manpower and hardware capability.). This includes minimum software and hardware data holding site requirement (e.g. ftp access and ESG authentication and authorization) and a skilled staff help desk. The AR5 archive is to be globally distributed with support for WG1, WG2, and WG3. Will there be a need for a central (or core) archive and what will it look like? Replication of holdings - disaster protection, a desire to have a replica of the core data archive on every continent, etc. Number of users and level of access – scientist, policy makers, economists, health officials, etc.

- Vendredi 27 mars

7

8 Orders of magnitude Climate models, centennial runs. Resolutions used Atmosphere 2.5° (280 Km) : 144 x 143 x 39 Ocean 2° (220 Km) : 180 x 149 x 31 Atm 2.5° - Ocean 2° : 20 GB/y, 300 ans  5,85 TB Atm 1.0° - Ocean 2° : 60 GB/y, 300 ans  17,5 TB Atm 0.5° - Ocean 0,5° : 400 GB/y, 30 ans  11,75 TB

- Vendredi 27 mars Global data amount Raw Data amount low bound  565 TB Raw Data amount high bound  1000 TB CMIP5 Distribution (25-50%)  ( ) ( ) TB Global Storage (Raw+Distributed)  TB LMDz 0.5° (50 Km)

- Vendredi 27 mars Management of data since years Mainly centralised, store on a SAN OpenDap access on Supercomputing Centre Basic system of data retrieval Access to raw data Security/Authentication/Restriction to data access : not an issue No on demand post-processing No metadata integration No support for high level database query

- Vendredi 27 mars Data management with Prodiguer Move the data a minimum, keep them close to supercomputing centres if possible  Data access protocol, strong links with computing centres When data needs to be moved do it quickly and with a minimum amount of human intervention  Management of storage resources, fast network Keep a track of what we got, particularly what is on deep storage  Metadata et data catalogues Exploiting of federation of sites  Grid middleware  Data grid ?