SDM Center February 2, 2005 Progress on MPI-IO Access to Mass Storage System Using a Storage Resource Manager Ekow J. Otoo, Arie Shoshani and Alex Sim.

Slides:



Advertisements
Similar presentations
SDM Center Coupling Parallel IO with Remote Data Access Ekow Otoo, Arie Shoshani, Doron Rotem, and Alex Sim Lawrence Berkeley National Lab.
Advertisements

Chapter 3. MPI MPI = Message Passing Interface Specification of message passing libraries for developers and users –Not a library by itself, but specifies.
1 OBJECTIVES To generate a web-based system enables to assemble model configurations. to submit these configurations on different.
1 SRM-Lite: overcoming the firewall barrier for large scale file replication Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory April, 2007.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Cache Coherent Distributed Shared Memory. Motivations Small processor count –SMP machines –Single shared memory with multiple processors interconnected.
1 CHEP 2003 Arie Shoshani Experience with Deploying Storage Resource Managers to Achieve Robust File replication Arie Shoshani Alex Sim Junmin Gu Scientific.
USING THE GLOBUS TOOLKIT This summary by: Asad Samar / CALTECH/CMS Ben Segal / CERN-IT FULL INFO AT:
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
Member of the ExperTeam Group Ralf Ratering Pallas GmbH Hermülheimer Straße Brühl, Germany
Cross Cluster Migration Remote access support Adianto Wibisono supervised by : Dr. Dick van Albada Kamil Iskra, M. Sc.
Grid IO APIs William Gropp Mathematics and Computer Science Division.
Managing Heterogeneous MPI Application Interoperation and Execution. From PVMPI to SNIPE based MPI_Connect() Graham E. Fagg*, Kevin S. London, Jack J.
ADLB Update Recent and Current Adventures with the Asynchronous Dynamic Load Balancing Library Rusty Lusk Mathematics and Computer Science Division Argonne.
Stream Handling Streams - means flow of data to and from program variables. - We declare the variables in our C++ for holding data temporarily in the memory.
Alok 1Northwestern University Access Patterns, Metadata, and Performance Alok Choudhary and Wei-Keng Liao Department of ECE,
Module 1: Introduction to C# Module 2: Variables and Data Types
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
Screen Snapshot Service Kurt Biery SiTracker Monitoring Meeting, 23-Jan-2007.
1 A Look at PVFS, a Parallel File System for Linux Talk originally given by Will Arensman and Anila Pillai.
Pooja Shetty Usha B Gowda.  Network File Systems (NFS)  Drawbacks of NFS  Parallel Virtual File Systems (PVFS)  PVFS components  PVFS application.
UNIX SVR4 COSC513 Zhaohui Chen Jiefei Huang. UNIX SVR4 UNIX system V release 4 is a major new release of the UNIX operating system, developed by AT&T.
SRM at Clemson Michael Fenn. What is a Storage Element? Provides grid-accessible storage space. Is accessible to applications running on OSG through either.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
A. Sim, CRD, L B N L 1 US CMS Workshop, Mar. 3, 2009 Berkeley Storage Manager (BeStMan) Alex Sim Scientific Data Management Research Group Computational.
CSCI 6962: Server-side Design and Programming Web Services.
A. Sim, CRD, L B N L 1 Oct. 23, 2008 BeStMan Extra Slides.
Parallel Interactive Computing with PyTrilinos and IPython Bill Spotz, SNL (Brian Granger, Tech-X Corporation) November 8, 2007 Trilinos Users Group Meeting.
 200 Total Points ◦ 74 Points Writing Programs ◦ 60 Points Tracing Algorithms and determining results ◦ 36 Points Short Answer ◦ 30 Points Multiple Choice.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
SUMA: A Scientific Metacomputer Cardinale, Yudith Figueira, Carlos Hernández, Emilio Baquero, Eduardo Berbín, Luis Bouza, Roberto Gamess, Eric García,
NoSQL Databases Oracle - Berkeley DB. Content A brief intro to NoSQL About Berkeley Db About our application.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
SciDAC All Hands Meeting, March 2-3, 2005 Northwestern University PIs:Alok Choudhary, Wei-keng Liao Graduate Students:Avery Ching, Kenin Coloma, Jianwei.
Opportunities in Parallel I/O for Scientific Data Management Rajeev Thakur and Rob Ross Mathematics and Computer Science Division Argonne National Laboratory.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Parallel and Grid I/O Infrastructure W. Gropp, R. Ross, R. Thakur Argonne National Lab A. Choudhary, W. Liao Northwestern University G. Abdulla, T. Eliassi-Rad.
1 Grid File Replication using Storage Resource Management Presented By Alex Sim Contributors: JLAB: Bryan Hess, Andy Kowalski Fermi: Don Petravick, Timur.
Computing Sciences Directorate, L B N L 1 CHEP 2003 Standards For Storage Resource Management BOF Co-Chair: Arie Shoshani * Co-Chair: Peter Kunszt ** *
1 SRM-Lite: overcoming the firewall barrier for data movement Arie Shoshani Alex Sim Viji Natarajan Lawrence Berkeley National Laboratory SDM Center All-Hands.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Active Storage Processing in Parallel File Systems Jarek Nieplocha Evan Felix Juan Piernas-Canovas SDM CENTER.
SDM Center Coupling Parallel IO to SRMs for Remote Data Access Ekow Otoo, Arie Shoshani and Alex Sim Lawrence Berkeley National Laboratory.
1 Use of SRM File Streaming by Gateway Alex Sim Arie Shoshani May 2008.
C O M P U T A T I O N A L R E S E A R C H D I V I S I O N SRM Basic/Advanced Spec Issues Arie Shoshani, Alex Sim, Junmin Gu Scientific Data Management.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
Project18 Communication Design + Parallelization Camilo A Silva BIOinformatics Summer 2008.
SDM Center Parallel I/O Storage Efficient Access Team.
Process Manager Specification Rusty Lusk 1/15/04.
3/12/2013Computer Engg, IIT(BHU)1 MPI-1. MESSAGE PASSING INTERFACE A message passing library specification Extended message-passing model Not a language.
CS 591x Overview of MPI-2. Major Features of MPI-2 Superset of MPI-1 Parallel IO (previously discussed) Standard Process Startup Dynamic Process Management.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
A. Sim, CRD, L B N L 1 SRM Collaboration Meeting, Sep , 2005 SRM v3.0 LBNL Implementation Status Report Scientific Data Management Research Group.
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
1 Scientific Data Management Group LBNL SRM related demos SC 2002 DemosDemos Robust File Replication of Massive Datasets on the Grid GridFTP-HPSS access.
An Introduction to GPFS
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
9/20/04Storage Resource Manager, Timur Perelmutov, Jon Bakken, Don Petravick, Fermilab 1 Storage Resource Manager Timur Perelmutov Jon Bakken Don Petravick.
A. Sim, CRD, L B N L 1 Production Data Management Workshop, Mar. 3, 2009 BeStMan and Xrootd Alex Sim Scientific Data Management Research Group Computational.
OSCAR Symposium – Quebec City, Canada – June 2008 Proposal for Modifications to the OSCAR Architecture to Address Challenges in Distributed System Management.
The Distributed Application Debugger (DAD)
Berkeley Storage Manager (BeStMan)
Network Load Balancing
MPI Message Passing Interface
CS 2308 Final Exam Review.
A Virtual Machine Monitor for Utilizing Non-dedicated Clusters
Presentation transcript:

SDM Center February 2, 2005 Progress on MPI-IO Access to Mass Storage System Using a Storage Resource Manager Ekow J. Otoo, Arie Shoshani and Alex Sim Lawrence Berkeley National Laboratory

SDM Center February 2, 2005 Objective To allow near-online transparent access to files on mass storage system (e.g., HPSS) from MPI applications. To enable existing applications to dynamically access files from MSS with little modification to the source code. To demonstrate the feasibility of the approach by running some applications on a Linux cluster using  PVFS as the local parallel file system and  HPSS as the mass storage system

SDM Center February 2, 2005 MPI Application on a Cluster Accessing an MSS ION Proxy + Svr CN Other MSS Parallel File System GridFTP CN – Compute Node ION – IO Node SRM-Proxy SRM-Svr – SRM Server Mass Storage System (MSS)

SDM Center February 2, 2005 MPI-IO-SRM Architecture Data Intensive Applications MPI-IOSRM PVFSGPFSUFSXFSOther BDBOther ADIO HPSS SAM Jasmin Castor High Level Access and Control Record Structured File Access Low Level File System Access pNetCDF HDF5

SDM Center February 2, 2005 Main Functions of MpiioSrm Library This is a package of library functions called libMpiioSrm.a  The functions callable from MPI applications are: 1.MPI_Srm_proxy_init(); 2.MPI_Srm_file_open(); [in place of MPI_File_open()] 3.MPI_Srm_file_close(); [in place of MPI_File_close()] 4.MPI_Srm_file_delete(); [in place of MPI_File_delete()] 5.MPI_Srm_proxy_close() Functions (2), (3) and (4) are the only ones that take file names as a parameter of all the MPI-IO library functions. These can be used to build complex execution models of an application.

SDM Center February 2, 2005 Dependent SRM Functions Functions equivalent to SRM commands are:  srm_abort(); srm_copy();  srm_get();srm_ls();  srm_mkdir();srm_ping();  srm_put();srm_release();  srm_remove();srm_status(); Parameters used in these functions are either derived from a configuration file “hrm.rc” or communicated through MPI_Info object.

SDM Center February 2, 2005 How to Use the Library Functions Requires inclusion of MpiioSrm.h for compilation and libMpiioSrm during linkage. After MPI_init() we call  MPI_Srm_proxy_init( MPI_Comm comm, int myrank, int srm_enabled, char *proxy_host, int proxy_rank, char *fileofurls, char fileofurl_fmt, char *srmuserid, char *hrmrcpath, MPI_Info *srminfo) ; The application then calls  MPI_Srm_file_open(), …   MPI_Srm_file_close(), etc. Finally it calls  MPI_Srm_proxy_close(MPI_Comm comm, int proxy_rank) before calling MPI_Finalize();

SDM Center February 2, 2005 Some Special Features of MpiioSrm Methods of specifying data sources and targets.  Simple file name – a string of characters of a directory path and basename  URLs  Indirectly through a file of either List of source and target URLs or List of SRM commands Modes for opening files:  In addition to the various MPI_MODE_* of the standard we add the mode MPI_MODE_PROPCHG_ON_CLOSE to propagate any changes on the local file to the MSS. An application can set modes through MPI_File_info_set() command in our case.

SDM Center February 2, 2005 Status of the MpiioSrm library - 1 The current status of the library includes  For SRM services, we have srm_abort(), srm_copy(), srm_ls(), srm_ping(), srm_status(), srm_release() and srm_remove().  For MPI-IO calling SRMs we have all five basic functions implemented, namely MPI_Srm_proxy_init(); MPI_Srm_file_open(); MPI_Srm_file_close(); MPI_Srm_file_delete(); MPI_Srm_proxy_close()

SDM Center February 2, 2005 Status of the MpiioSrm library - 2 Test Programs  The current status includes the following test programs: TestSRM_S1RC – Static Single Request at a time. TestSRM_SRSC – Static Requests with Single file processing at a time. TestSRM_DRSOC – Dynamic Requests with Sequential Processing Order. TestSRM_DRAOC – Dynamic Requests with Arbitrary Processing order.

SDM Center February 2, 2005 Looking to the Future Short Term  Add srm_put() and srm_get() and srm_mkdir()  Improve method of communicating prefetched file with processes using MPI’s Remote Memory Access.  Release for general use. Long Term  Address scalability, fault tolerance issues and PVFS2 – use of multiple SRM servers.  Define XML schema for input file formats  GUI for: Preparing input data files for applics Monitor files fetched and processed by applications through MPI’s Dynamic Process Management functions