Joseph Antony, Andrew Howard, Jason Andrade, Ben Evans, Claire Trenham, Jingbo Wang Production Petascale Climate Data Replication at.

Slides:



Advertisements
Similar presentations
Jens G Jensen Atlas Petabyte store Supporting Multiple Interfaces to Mass Storage Providing Tape and Mass Storage to Diverse Scientific Communities.
Advertisements

A. Sim, CRD, L B N L 1 ANI and Magellan Launch, Nov. 18, 2009 Climate 100: Scaling the Earth System Grid to 100Gbps Networks Alex Sim, CRD, LBNL Dean N.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Preservation and Long Term Access of Data at the World Data Centre for Climate Frank Toussaint N.P. Drakenberg, H. Höck, M. Lautenschlager, H. Luthardt,
Christopher Paolini Computational Science Research Center San Diego State University 100G and Beyond Workshop: Ultra High Performance Networking in California.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Computing and Data Infrastructure for Large-Scale Science Deploying Production Grids: NASA’s IPG and DOE’s Science Grid William E. Johnston
Mike Smorul Saurabh Channan Digital Preservation and Archiving at the Institute for Advanced Computer Studies University of Maryland, College Park.
CLIMATE SCIENTISTS’ BIG CHALLENGE: REPRODUCIBILITY USING BIG DATA Kyo Lee, Chris Mattmann, and RCMES team Jet Propulsion Laboratory (JPL), Caltech.
Large-Scale Data Collection Metadata Management at the National Computational Infrastructure (NCI) Jingbo Wang 1, Irina Bastrakova.
M. Stockhause et al. Martina Stockhause, Michael Lautenschlager, Frank Toussaint Deutsches Klimarechenzentrum (DKRZ) World Data Centre for Climate (WDCC)
Z EGU Integration of external metadata into the Earth System Grid Federation (ESGF) K. Berger 1, G. Levavasseur 2, M. Stockhause 1, and M. Lautenschlager.
TeraGrid Information Services December 1, 2006 JP Navarro GIG Software Integration.
Introductionto Networking Basics By Avinash Kulkarni.
Presenter: Dipesh Gautam.  Introduction  Why Data Grid?  High Level View  Design Considerations  Data Grid Services  Topology  Grids and Cloud.
Topology Level: Prep-I Topic: Computer Networks Ms. Fouzia Yousuf October 25, 2010.
Presented by The Earth System Grid: Turning Climate Datasets into Community Resources David E. Bernholdt, ORNL on behalf of the Earth System Grid team.
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Crystal25 Hunter Valley, Australia, 11 April 2007 Crystal25 Hunter Valley, Australia, 11 April 2007 JAINIS (JCU and Indiana Instrument Services): A Grid.
High Performance GridFTP Transport of Earth System Grid (ESG) Data 1 Center for Enabling Distributed Petascale Science.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks G. Quigley, B. Coghlan, J. Ryan (TCD). A.
Data Publication and Quality Control Procedure for CMIP5 / IPCC-AR5 Data WDC Climate / DKRZ:
- EGU 2010 ESSI May Building on the CMIP5 effort to prepare next steps : integrate community related effort in the every day workflow to.
Authors: Ronnie Julio Cole David
- Vendredi 27 mars PRODIGUER un nœud de distribution des données CMIP5 GIEC/IPCC Sébastien Denvil Pôle de Modélisation, IPSL.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Chapter2 Networking Fundamentals
Data Transfers in the ALCF Robert Scott Technical Support Analyst Argonne Leadership Computing Facility.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
7. Grid Computing Systems and Resource Management
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
1 Overall Architectural Design of the Earth System Grid.
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
LLNL’s Data Center and Interoperable Services 5 th Annual ESGF Face-to-Face Conference ESGF 2015 Monterey, CA, USA Dean N. Williams, Tony Hoang, Cameron.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
1 Summary. 2 ESG-CET Purpose and Objectives Purpose  Provide climate researchers worldwide with access to data, information, models, analysis tools,
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Globus and ESGF Rachana Ananthakrishnan University of Chicago
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
LLNL-PRES-XXXXXX This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under contract DE-AC52-07NA27344.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI How to integrate portals with the EGI monitoring system Dusan Vudragovic.
Nanbor Wang, Balamurali Ananthan Tech-X Corporation Gerald Gieraltowski, Edward May, Alexandre Vaniachine Argonne National Laboratory 2. ARCHITECTURE GSIMF:
Data and storage services on the NGS.
Support to scientific research on seasonal-to-decadal climate and air quality modelling Pierre-Antoine Bretonnière Francesco Benincasa IC3-BSC - Spain.
Jost von Hardenberg ISAC-CNR, Torino, Italy with Paolo Davini, Susanna Corti, and many others EUDAT User Forum, Rome,Italy 3-4 February, 2016.
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
AOLI 2015 The NMME Experience: A Research Community Archive Lessons learned from Climate Model data archive and use AOLI Meeting 2015 Eric Nienhouse NCAR.
Nci.org.au © National Computational Infrastructure 2016 Virtual Laboratories in Australia Lesley Wyborn (NCI) With contributions from:
Weigel, Berger, Kindermann, Lautenschlager EGU Versioning for CMIP6 in the Earth System Grid Federation Data preparation Initial registration.
A Quick Tour of the NOAA Environmental Software Infrastructure and Interoperability Group Cecelia DeLuca Dr. Robert Detrick visit March 28, 2012
1 This slide indicated the continuous cycle of creating raw data or derived data based on collections of existing data. Identify components that could.
Intentions and Goals Comparison of core documents from DFIG and Publishing Workflow IG show that there is much overlap despite different starting points.
WP18, High-speed data recording Krzysztof Wrona, European XFEL
AP7/AP8: Long-Term Archival of CMIP6 Data
Dag Toppe Larsen UiB/CERN CERN,
Dag Toppe Larsen UiB/CERN CERN,
More Data Management and Services
Updates on HPC and Data management at NCI
Software infrastructure for a National Research Platform
ESnet and Science DMZs: an update from the US
Staying afloat in the sensor data deluge
CMIP6 use case and adoption of RDA outputs
Introduction to RDS Datasets
RDA uptake activities and plans: ESGF
Digital Object Management for ENES: Challenges and Opportunities
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Joseph Antony, Andrew Howard, Jason Andrade, Ben Evans, Claire Trenham, Jingbo Wang Production Petascale Climate Data Replication at NCI – Lustre and our engagement with the Earth Systems Grid Federation (ESGF)

nci.org.au MOTIVATION

nci.org.au International Climate Change Research – The CMIP projects The UN’s International Panel on Climate Change (IPCC) prepares an intergovernmental assessment report every 6 years This effort requires significant scientific and HPC/HPD resources to back it The most recent of these activities was the Coupled Model Intercomparison Project 5 (CMIP5) The NCI is a major data node within the ESGF federation In this talk I will share with you a ‘view from the coalface’, replicating ~2PB of data

nci.org.au

CMIP DATA VOLUMES

nci.org.au CMIP1 thru CMIP5 Data Volumes Taken from Dean Williams’ ESGF Internet2 presentation, 2014

nci.org.au ESGF NODE ARCHITECTURE

nci.org.au The ESGF Data Archival and Retrieval System The ESGF is a federated peer-to-peer international data archival and retrieval system Incorporates single- sign-on for end-users It has publication and version management tools Supports data aggregations and can notify users if datasets have been modified

nci.org.au THE END-USER PERSPECTIVE

nci.org.au The Last-Mile Problem … Data is too large to move onto desktop for analysis – CMIP3 to CMIP5 Users want versioned, curated data to be able to jump right into scientific analysis At NCI – An integrated eco-system exists for data- intensive science Data Repositories Virtual Laboratories – The ICNWG effort to solve the ‘Last Mile Problem’ for networking

nci.org.au ICNWG Activities

nci.org.au Okay … so where’s Lustre in all of this you ask?

nci.org.au Okay … so where’s Lustre in all of this you ask? We use Lustre as our distributed filesystem for a set of dedicated WAN data transfer nodes (DTNs)

nci.org.au Okay … so where’s Lustre in all of this you ask? We use Lustre as our distributed filesystem for a set of dedicated WAN data transfer nodes (DTNs) But first a detour …

nci.org.au Courtesy Eli Dart, ESnet 1Gbps == 125 MB/sec

nci.org.au Courtesy Eli Dart, ESnet

nci.org.au Courtesy Eli Dart, ESnet

nci.org.au Courtesy Eli Dart, ESnet

nci.org.au Courtesy Eli Dart, ESnet

nci.org.au

AARNet International Links

nci.org.au NCI’s DTN Nodes

nci.org.au CBR-SYD and onto the CONUS via SXtransport

nci.org.au Cable StationNetwork Segment SXtransport – Physical Layout

nci.org.au SXtransport – Logical Network Layout

nci.org.au What are some of the world’s longest submarine cables you ask? 39,000 Km of submarine fibre

nci.org.au What are some of the world’s longest submarine cables you ask? 39,000 Km of submarine fibre 28,900 Km of submarine fibre 1,600 Km of terrestrial fibre

nci.org.au Networking Topology for Data Replication Courtesy Mary Hester, ESnet

nci.org.au Initial Transfer Rates from NCI Graph shows the data rate vs. the volume of data transferred Different lines in the graph represent how many data streams were required to obtain the given performance. The results of the graph indicate that it is possible to get a line-rate of 1GB/s (8Gbps) between Australia and the United States, however, it requires configuring transfers to run more than 100 parallel streams

nci.org.au Data replication and Science DMZs Currently we’ve replicated ~1.5PB Working on improving these rates by employing a Science DMZ model and dedicated data transfer nodes

nci.org.au Globus Online Globus Online is a hosted data-transfer-as- a-service offering, run by the University of Chicago It makes the job of large data transfers easy for both instrument owners and end-users

nci.org.au Globus Online Architecture

nci.org.au

Using Dedicated DTNs – January 2015

nci.org.au Using Dedicated DTNs – March 2015

nci.org.au State of the Union Numbers from the ICNWG Consortium

nci.org.au Conclusion Non-trivial to get various ducks lined-up – 10GigE WAN networking – Mellanox tuning work for 10GigE Ethernet and 56Gbp FDR – Being NUMA aware is critical for the GridFTP daemon!

nci.org.au THE END

nci.org.au VERIFIED, CURATED SCIENTIFIC DATASETS

nci.org.au Centralized Quality Control for Data Processing Multi-layered QC – Initial Level 1 QC done at data nodes – DKRZ performs L2 QC – Further metadata and variable checking is done to get to L3 QC At every step, end-users can see the QC Level for their data Replicated data has passed QC Level 3 and receives a DOI