1 The SciDAC Scientific Data Management Center: Infrastructure and Results Arie Shoshani Lawrence Berkeley National Laboratory SC 2004 November, 2004.

Slides:



Advertisements
Similar presentations
1 Senn, Information Technology, 3 rd Edition © 2004 Pearson Prentice Hall James A. Senns Information Technology, 3 rd Edition Chapter 7 Enterprise Databases.
Advertisements

1 Towards an Open Service Framework for Cloud-based Knowledge Discovery Domenico Talia ICAR-CNR & UNIVERSITY OF CALABRIA, Italy Cloud.
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
11 Application of CSF4 in Avian Flu Grid: Meta-scheduler CSF4. Lab of Grid Computing and Network Security Jilin University, Changchun, China Hongliang.
© 2006 Open Grid Forum GGF18, 13th September 2006 OGSA Data Architecture Scenarios Dave Berry & Stephen Davey.
DCV: A Causality Detection Approach for Large- scale Dynamic Collaboration Environments Jiang-Ming Yang Microsoft Research Asia Ning Gu, Qi-Wei Zhang,
Designing Services for Grid-based Knowledge Discovery A. Congiusta, A. Pugliese, Domenico Talia, P. Trunfio DEIS University of Calabria ITALY
So far Binary numbers Logic gates Digital circuits process data using gates – Half and full adder Data storage – Electronic memory – Magnetic memory –
Chapter 1 Introduction Copyright © Operating Systems, by Dhananjay Dhamdhere Copyright © Introduction Abstract Views of an Operating System.
World Health Organization
Database Systems: Design, Implementation, and Management
Computer Literacy BASICS
The HDF Group Parallel HDF5 Developments 1 Copyright © 2010 The HDF Group. All Rights Reserved Quincey Koziol The HDF Group
The Platform as a Service Model for Networking Eric Keller, Jennifer Rexford Princeton University INM/WREN 2010.
Executional Architecture
Global Analysis and Distributed Systems Software Architecture Lecture # 5-6.
Systems Analysis and Design in a Changing World, Fifth Edition
University of Chicago Department of Energy The Parallel and Grid I/O Perspective MPI, MPI-IO, NetCDF, and HDF5 are in common use Multi TB datasets also.
1 SRM-Lite: overcoming the firewall barrier for large scale file replication Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory April, 2007.
1 CHEP 2003 Arie Shoshani Experience with Deploying Storage Resource Managers to Achieve Robust File replication Arie Shoshani Alex Sim Junmin Gu Scientific.
SDM center Questions – Dave Nelson What kind of processing / queries / searches biologists do over microarray data? –Range query on a spot? –Range query.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Toni Saarinen, Tite4 Tomi Ruuska, Tite4 Earth System Grid - ESG.
Astrophysics, Biology, Climate, Combustion, Fusion, Nanoscience Working Group on Simulation-Driven Applications 10 CS, 10 Sim, 1 VR.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Advanced Topics: MapReduce ECE 454 Computer Systems Programming Topics: Reductions Implemented in Distributed Frameworks Distributed Key-Value Stores Hadoop.
Scientific Data Management (SDM)
Presented by The Earth System Grid: Turning Climate Datasets into Community Resources David E. Bernholdt, ORNL on behalf of the Earth System Grid team.
DataGrid Middleware: Enabling Big Science on Big Data One of the most demanding and important challenges that we face as we attempt to construct the distributed.
1 Scientific Data Management Center DOE Laboratories: ANL: Rob Ross LBNL:Doron Rotem LLNL:Chandrika Kamath ORNL: Nagiza Samatova.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
1 Arie Shoshani, LBNL SDM center Scientific Data Management Center(SDM-ISIC) Arie Shoshani Computing Sciences Directorate Lawrence Berkeley National Laboratory.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Opportunities in Parallel I/O for Scientific Data Management Rajeev Thakur and Rob Ross Mathematics and Computer Science Division Argonne National Laboratory.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Computer Science Research and Development Department Computing Sciences Directorate, L B N L 1 Storage Management and Data Mining in High Energy Physics.
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
SDM Center’s Data Mining & Analysis SDM Center Parallel Statistical Analysis with RScaLAPACK Parallel, Remote & Interactive Visual Analysis with ASPECT.
Intergrid KoM Santander 22 june, 2006 E-Infraestructure shared between Europe and Latin America José Manuel Gutiérrez
1 Arie Shoshani, LBNL SDM center Scientific Data Management Center (Integrated Software Infrastructure Center – ISIC) Arie Shoshani All Hands Meeting March.
The Earth System Grid: A Visualisation Solution Gary Strand.
1 Scientific Data Management Center(ISIC) contains extensive publication list.
Web Portal Design Workshop, Boulder (CO), Jan 2003 Luca Cinquini (NCAR, ESG) The ESG and NCAR Web Portals Luca Cinquini NCAR, ESG Outline: 1.ESG Data Services.
The Earth System Grid (ESG) Computer Science and Technologies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
Presented by Scientific Data Management Center Nagiza F. Samatova Network and Cluster Computing Computer Sciences and Mathematics Division.
F. Douglas Swesty, DOE Office of Science Data Management Workshop, SLAC March Data Management Needs for Nuclear-Astrophysical Simulation at the Ultrascale.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
1 SRM-Lite: overcoming the firewall barrier for data movement Arie Shoshani Alex Sim Viji Natarajan Lawrence Berkeley National Laboratory SDM Center All-Hands.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
1 Overall Architectural Design of the Earth System Grid.
Computing Sciences Directorate, L B N L 1 SC 2003 Storage Resource Managers: Essential Components for the Grid Arie Shoshani Staff: Alex Sim, Junmin Gu,
Supercomputing 2006 Scientific Data Management Center Lead Institution: LBNL; PI: Arie Shoshani Laboratories: ANL, ORNL, LBNL, LLNL, PNNL Universities:
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
PPDG meeting, July 2000 Interfacing the Storage Resource Broker (SRB) to the Hierarchical Resource Manager (HRM) Arie Shoshani, Alex Sim (LBNL) Reagan.
Climate-SDM (1) Climate analysis use case –Described by: Marcia Branstetter Use case description –Data obtained from ESG –Using a sequence steps in analysis,
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
Workflow Management Concepts and Requirements For Scientific Applications.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
1 Scientific Data Management Group LBNL SRM related demos SC 2002 DemosDemos Robust File Replication of Massive Datasets on the Grid GridFTP-HPSS access.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
The Earth System Grid: A Visualisation Solution
Scientific Data Management contains extensive publication list
SDM workshop Strawman report History and Progress and Goal.
TeraScale Supernova Initiative
Presentation transcript:

1 The SciDAC Scientific Data Management Center: Infrastructure and Results Arie Shoshani Lawrence Berkeley National Laboratory SC 2004 November, 2004

2 A Typical SDM Scenario Control Flow Layer Applications & Software Tools Layer I/O System Layer Storage & Network Resouces Layer Flow Tier Work Tier + Data Mover Simulation Program Parallel R Post Processing Terascale Browser Task A: Generate Time-Steps Task B: Move TS Task C: Analyze TS Task D: Visualize TS Parallel NetCDF PVFSLN HDF5 Libraries SRM

3 Dealing with Large Data volumes: Three Stages of Data Handling Generate data Run simulations, dump data Capture metadata Post-processing of data Process all data, produce reduced datasets, summaries Generate indexes Analyze data Interested in subset of the data, search over large amounts of data, produce relatively little data (for analysis, vis, …) Generation Post-processing Analysis Input Output Computation Low High High High Med-High High Med-High Low Low-Med Stage Data

4 I/O is the predicted bottleneck Source: Celeste Matarazzo Main reason: data transfer rates to disk and tape devices have not kept pace with computational capacity

5 Data Generation – technology areas Parallel I/O writes – to disk farms Does technology scale? Reliability in face of disk failures Dealing with various file formats (NetCDF, HDF, AMR, unstructured meshes, …) Parallel Archiving – to tertiary storage Is tape striping cost effective? Reorganizing data before archiving to match predicted access patterns Dynamic monitoring of simulation progress Tools to automate workflow Compression – is overhead prohibitive?

6 Post Processing – technology areas Parallel I/O reads – from disk farms Data clustering to minimize arm movement Parallel read synchronization Does it scale linearly? Reading from archives – tertiary storage Minimize tape mounts Does tape striping help? Whats after tape? Large disk farm archives? Feed large volumes data into machine Competes with write I/O Need fat I/O channels, parallel I/O

7 Analysis – technology areas Dynamic hot clustering of data from archive Based on repeated use (caching & replication) Takes advantage of data sharing Indexing over data values Indexes need to scale: Linear search over billion of data objects Search over combinations of multiple data measures per mesh point Take advantage of append-only data Parallel indexing methods Analysis result streaming On the fly monitoring and visualization Suspend-Resume capabilities, clean abort

8 SDM Technology: Layering of Components Hardware, OS, and MSS (HPSS) Scientific Process Automation (Workflow) Layer Data Mining & Analysis Layer Storage Efficient Access Layer

9 Data Generation Workflow Design and Execution Scientific Process Automation Layer OS, Hardware (Disks, Mass Store) Storage Efficient Access Layer Data Mining and Analysis Layer Simulation Run Parallel netCDF MPI-IOPVFS2

10 Parallel NetCDF v.s. NetCDF (ANL+NWU) Slow and cumbersome Data shipping I/O bottleneck Memory requirement netCDF Parallel File System P2P3P0P1 Parallel File System Parallel netCDF P2P3P0P1 Programming Convenience Perform I/O cooperatively or collectively Potential parallel I/O optimizations for better performance

11 Parallel NetCDF Library Overview User level library Accept parallel requests in netCDF I/O patterns Parallel I/O through MPI- IO to underlying file system and storage Good level of abstraction for portability and optimization opportunities

12 Parallel netCDF Performance

13 Robust Multi-file Replication Problem: move thousands of files robustly Takes many hours Need error recovery Mass storage systems failures Network failures Solution: Use Storage Resource Managers (SRMs) Problem: too slow Solution: Use parallel streams Use concurrent transfers Use large FTP windows Pre-stage files from MSS NCAR Anywhere LBNL Disk Cache Disk Cache SRM-COPY (thousands of files) SRM-GET (one file at a time) DataMover SRM (performs writes) SRM (performs reads) GridFTP GET (pull mode) stage files archive files Network transfer Get list of files MSS

14 File tracking shows recovery from transient failures Total: 45 GBs

15 Tomcat servlet engine Tomcat servlet engine MCS Metadata Cataloguing Services MCS Metadata Cataloguing Services RLS Replica Location Services RLS Replica Location Services SOAP RMI MyProxy server MyProxy server MCS client RLS client MyProxy client GRAM gatekeeper GRAM gatekeeper CAS Community Authorization Services CAS Community Authorization Services CAS client NCAR-MSS Mass Storage System HPSS High Performance Storage System HPSS High Performance Storage System DRM Storage Resource Management DRM Storage Resource Management HRM Storage Resource Management HRM Storage Resource Management HRM Storage Resource Management HRM Storage Resource Management HRM Storage Resource Management HRM Storage Resource Management gridFTP server gridFTP server gridFTP server gridFTP server gridFTP server gridFTP server gridFTP server gridFTP server openDAPg server openDAPg server gridFTP Striped server gridFTP Striped server LBNL LLNL USC-ISI NCAR ORNL ANL DRM Storage Resource Management DRM Storage Resource Management disk Earth Science Grid

16 Layering of Components Hardware, OS, and MSS (HPSS) Scientific Process Automation (Workflow) Layer Data Mining & Analysis Layer Storage Efficient Access Layer

17 RScaLAPACK (ORNL) Through RScaLAPACK we provide a simple intuitive R interfaces to ScaLAPACK routines. The package does not need the user to worry about setting up the parallel environment and distribution of data prior to ScaLAPACK function call. RScaLAPACK is developed as an add-on package to R. Significant speed gain is observed in some function execution. Submitted to CRAN ( a network of 27 www sites across 15 countries holding R distribution) in March 2003

18 (1) Parallel Agent (PA): Orchestrates entire parallel execution as per users request. (2) Spawned Processes: Actual parallel execution of the requested function. RScaLAPACK Architecture

19 RScaLAPACK Benchmarks

20 Fastbit – bitmap Indexing Technology (LBNL) Search over large spatio-temporal data Combustion simulation: 1000x1000x1000 mesh with 100s of chemical species over 1000s of time steps Supernova simulation: 1000x1000x1000 mesh with 10s of variables per cell over 1000s of time steps Common searches are partial range queries Temperature > 1000 AND pressure > 10 6 HO 2 > AND HO 2 > Features Search time proportional to number of hits Index generation linear with data values (require read-once only)

21 FastBit-Based Multi-Attribute Region Finding is Theoretically Optimal On 3D data with over 110 million points, region finding takes less than 2 seconds Flame Front discovery (range conditions for multiple measures) in a combustion simulation (Sandia) Time required to identify regions in 3D Supernova simulation (LBNL)

22 Feature Selection and Extraction: Using Generic Analysis Tools (LLNL) Comparing Climate simulation to experiment data Used PCA and ICA technology for accurate Climate signal separation

23 Layering of Components Hardware, OS, and MSS (HPSS) Scientific Process Automation (Workflow) Layer Data Mining & Analysis Layer Storage Efficient Access Layer

24 TSI Workflow Example In conjunction with John Blondin, NCSU Automate data generation, transfer and visualization of a large-scale simulation at ORNL

25 TSI Workflow Example In conjunction with John Blondin, NCSU Automate data generation, transfer and visualization of a large-scale simulation at ORNL Submit Job to Cray at ORNL Check whether a time slice is finished Aggregate all into One large File - Save to HPSS Split it into 22 Files and store them in XRaid Notify Head Node at NC State Head Node submit scheduling to SGE SGE schedule the transfer for 22 Nodes Node Retrieve File from XRaid Node-0 Node-1 ….. Node-21 Yes Start Ensight to generate Video Files at Head Node ORNL NCSU

26 Using the Scientific Workflow Tool (Kepler) Emphasizing Dataflow (SDSC, NCSU, LLNL) Automate data generation, transfer and visualization of a large-scale simulation at ORNL

27 TSI Workflow Example with Doug Swesty and Eric Myra, Stony Brook Automate the transfer of large-scale simulation data between NERSC and Stony Brook

28 Using the Scientific Workflow Tool

29 Collaborating Application Scientists Matt Coleman - LLNL (Biology) Tony Mezzacappa – ORNL (Astrophysics) Ben Santer – LLNL John Drake - ORNL (Climate) Doug Olson - LBNL, Wei-Ming Zhang – Kent (HENP) Wendy Koegler, Jacqueline Chen – Sandia Lab (Combustion) Mike Papka - ANL (Astrophysics Vis) Mike Zingale – U of Chicago (Astrophysics) John Michalakes – NCAR (Climate) Keith Burrell - General Atomics (Fusion)

30 Re-apply technology to new applications Parallel NetCDF Astrophysics Climate Parallel VTK Astrophysics Climate Compressed bitmaps HENP Combustion Astrophysics Storage Resource Managers (MSS access) HENP Climate Astrophysics Feature Selection Climate Fusion Scientific Workflow Biology Astrophysics (planned)

31 Summary Focus: getting technology into the hands of scientists Integrated framework Storage Efficient Access technology Data Mining and Analysis tools Scientific Process (workflow) Automation Technology migration to new applications SDM Framework facilitates integration, generalization, and reuse of SDM technology