PPDG meeting, July 2000 Interfacing the Storage Resource Broker (SRB) to the Hierarchical Resource Manager (HRM) Arie Shoshani, Alex Sim (LBNL) Reagan.

Slides:



Advertisements
Similar presentations
NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE SAN DIEGO SUPERCOMPUTER CENTER Particle Physics Data Grid PPDG Data Handling System Reagan.
Advertisements

1 SRM-Lite: overcoming the firewall barrier for large scale file replication Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory April, 2007.
Computing Sciences Directorate, L B N L 1 CHEP 2003 Storage Resource Management In the Grid Environment Alex Sim Junmin Gu Arie Shoshani Scientific Data.
1 CHEP 2003 Arie Shoshani Experience with Deploying Storage Resource Managers to Achieve Robust File replication Arie Shoshani Alex Sim Junmin Gu Scientific.
Aug Arie Shoshani Particle Physics Data Grid Request Management working group.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
A Grid Resource Broker Supporting Advance Reservations and Benchmark- Based Resource Selection Erik Elmroth and Johan Tordsson Reporter : S.Y.Chen.
Applying Data Grids to Support Distributed Data Management Storage Resource Broker Reagan W. Moore Ian Fisk Bing Zhu University of California, San Diego.
Page 1 Processes and Threads Chapter Processes 2.2 Threads 2.3 Interprocess communication 2.4 Classical IPC problems 2.5 Scheduling.
Based on the slides supporting the text
I/O Hardware n Incredible variety of I/O devices n Common concepts: – Port – connection point to the computer – Bus (daisy chain or shared direct access)
Physical Design CS 543 – Data Warehousing. CS Data Warehousing (Sp ) - Asim LUMS2 Physical Design Steps 1. Develop standards 2.
12: IO Systems1 I/O SYSTEMS This Chapter is About Processor I/O Interfaces: Connections exist between processors, memory, and IO devices. The OS must manage.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
1 Input/Output Chapter 3 TOPICS Principles of I/O hardware Principles of I/O software I/O software layers Disks Clocks Reference: Operating Systems Design.
1 I/O Management in Representative Operating Systems.
1 Today I/O Systems Storage. 2 I/O Devices Many different kinds of I/O devices Software that controls them: device drivers.
PRASHANTHI NARAYAN NETTEM.
STACS STACS: Storage Access Coordination of Tertiary Storage for High Energy Physics Applications Arie Shoshani, Alex Sim, John Wu, Luis Bernardo*, Henrik.
Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
I/O Tanenbaum, ch. 5 p. 329 – 427 Silberschatz, ch. 13 p
I/O Systems ◦ Operating Systems ◦ CS550. Note:  Based on Operating Systems Concepts by Silberschatz, Galvin, and Gagne  Strongly recommended to read.
Disk Access. DISK STRUCTURE Sector: Smallest unit of data transfer from/to disk; 512B 2/4/8 adjacent sectors transferred together: Blocks Read/write heads.
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
1 I/O Management and Disk Scheduling Chapter Categories of I/O Devices Human readable Used to communicate with the user Printers Video display terminals.
CSC 322 Operating Systems Concepts Lecture - 25: by Ahmed Mumtaz Mustehsan Special Thanks To: Tanenbaum, Modern Operating Systems 3 e, (c) 2008 Prentice-Hall,
Hardware Definitions –Port: Point of connection –Bus: Interface Daisy Chain (A=>B=>…=>X) Shared Direct Device Access –Controller: Device Electronics –Registers:
Recall: Three I/O Methods Synchronous: Wait for I/O operation to complete. Asynchronous: Post I/O request and switch to other work. DMA (Direct Memory.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
File and Object Replication in Data Grids Chin-Yi Tsai.
INFSO-RI Enabling Grids for E-sciencE DAGs with data placement nodes: the “shish-kebab” jobs Francesco Prelz Enzo Martelli INFN.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Computer Science Research and Development Department Computing Sciences Directorate, L B N L 1 Storage Management and Data Mining in High Energy Physics.
Device Drivers CPU I/O Interface Device Driver DEVICECONTROL OPERATIONSDATA TRANSFER OPERATIONS Disk Seek to Sector, Track, Cyl. Seek Home Position.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Bulk Data Movement: Components and Architectural Diagram Alex Sim Arie Shoshani LBNL April 2009.
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
1 Meeting Location: LBNL Sept 18, 2003 The functionality of a Replica Registration Service Attendees Michael Haddox-Schatz, JLAB Ann Chervenak, USC/ISI.
SDM-Center talk Optimizing shared access to tertiary storage Arie Shoshani Alex Sim Alex Sim July 10, 2001 Scientific Data Management Group Computing Science.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
1 Grid File Replication using Storage Resource Management Presented By Alex Sim Contributors: JLAB: Bryan Hess, Andy Kowalski Fermi: Don Petravick, Timur.
Computing Sciences Directorate, L B N L 1 CHEP 2003 Standards For Storage Resource Management BOF Co-Chair: Arie Shoshani * Co-Chair: Peter Kunszt ** *
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
January 26, 2003Eric Hjort HRMs in STAR Eric Hjort, LBNL (STAR/PPDG Collaborations)
Grand Challenge in MDC2 D. Olson, LBNL 31 Jan 1999 STAR Collaboration Meeting
Silberschatz, Galvin and Gagne  Operating System Concepts Six Step Process to Perform DMA Transfer.
I/O, Devices & Device Drivers I/O subsystem It is the responsibility of the I/O subsystem of an OS to schedule, manage, & control the interactions between.
SDM Center Coupling Parallel IO to SRMs for Remote Data Access Ekow Otoo, Arie Shoshani and Alex Sim Lawrence Berkeley National Laboratory.
1 Use of SRM File Streaming by Gateway Alex Sim Arie Shoshani May 2008.
11.1 Silberschatz, Galvin and Gagne ©2005 Operating System Principles 11.5 Free-Space Management Bit vector (n blocks) … 012n-1 bit[i] =  1  block[i]
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
4P13 Week 9 Talking Points
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 7 – Buffer Management.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
FILE SYSTEM IMPLEMENTATION 1. 2 File-System Structure File structure Logical storage unit Collection of related information File system resides on secondary.
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
SRM-iRODS Interface Development WeiLong UENG Academia Sinica Grid Computing 1.
1 Scientific Data Management Group LBNL SRM related demos SC 2002 DemosDemos Robust File Replication of Massive Datasets on the Grid GridFTP-HPSS access.
CASTOR: possible evolution into the LHC era
Processes and threads.
Module 11: File Structure
Operating System I/O System Monday, August 11, 2008.
CS703 - Advanced Operating Systems
Overview Continuation from Monday (File system implementation)
Chapter 2: Operating-System Structures
Chapter 2: Operating-System Structures
Presentation transcript:

PPDG meeting, July 2000 Interfacing the Storage Resource Broker (SRB) to the Hierarchical Resource Manager (HRM) Arie Shoshani, Alex Sim (LBNL) Reagan Moore, Bing Zhu (SDSC) PPDG meeting July 13-14, 2000

PPDG meeting, July 2000 Outline HRM role in the Data Grid architecture HRM API description HRM - HPSS system functionality SRB Client calls SRB-HRM architecture Status

PPDG meeting, July 2000 HRM role in the Data Grid architecture The class of Storage Resource Managers (SRMs) includes: —HRM: for managing the access to tape resources may or may not have a disk cache functionality generic but needs to be specialize for specific mass storage systems e.g. HRM-HPSS, HRM-Enstore,... —DRM: for managing disk resources functionality generic but needs to be specialize for specific disk systems e.g. DRM-FileSystem, DRM-DPSS, …

PPDG meeting, July 2000 Functionality Examples HRM functionality may include : -- queuing of file transfer requests (kind of a reservation) -- reordering of requests to optimize PFTP order (ordered by files on the same tape) -- Monitoring progress and error messages -- rescheduling transfers that failed DRM functionality may include: -- keeping tracks of files in cache -- managing space reservations -- making decisions on which files to remove when space is needed -- optimizing cache use - sharing files requested by multiple clients -- enforce local policy for cache use HRM role in the Data Grid architecture

PPDG meeting, July 2000 HRMs & DRMs in a Grid Architecture tape system client Request Manager HRM Disk Cache client Request Manager DRM Disk Cache Site 2Site 1 Advantages: 1. Each SRM has global view of all requests to it. 2. SRMs support reservation requests for space and file transfers 3. Can develop uniform error handling codes for all resources 4. Can ask SRMs to queue and monitor requests. 5. Main implication: need APIs to invoke SRMs, and a way for the Grid to pass requests to them. DRM

PPDG meeting, July 2000 HRM API description API Functionality —Request to transfer a file to destination disk A blocking call —Request to stage a file to HRM disk A non-blocking call when HRM disk exists —Request status/time_estimate How long before file request will be processed —Request to abort a file transfer or stage In case file no longer needed —Release a file After file was moved to destination optional to improve system efficiency —Call back when file is staged

PPDG meeting, July 2000 Examples of call sequences actual transferwait for turn Release file transfer request Time Actual transfer Status, turn reached stage request Time turn reached transfer request status actual transfer transfer request Time turn reached stage request status Release file Release file abort

PPDG meeting, July 2000 CORBA IDL for HRM

PPDG meeting, July 2000 HRM-HPSS system functionality All transfers go through HRM disk —reasons: flexibility of pre-staging —disk is sufficiently cheap for a large cache —opportunity to optimize for same file requests Functionality —queuing file transfers —file queue management —File clustering parameter —Transfer rate estimation —Query estimation - total time —Error handling

PPDG meeting, July 2000 SRB Client calls Sget —blocking call (wait till file is transferred) —should only be issued if space on client disk is allocated Sstage —non-blocking call (returns time estimate) —requires a server if want to be notified —can issue status to find out if file was cached Sstatus —returns time estimate for file to be staged to HRM’s disk Sabort —cancel this file request

PPDG meeting, July 2000 PPDG Grid Architecture tape system client Request Manager HRM Disk Cache client Request Manager DRM Disk Cache Site 2Site 1 DRM Event-file index Logical query Replica catalog Network Weather Service Event-file index Replica catalog Network Weather Service File transfer requests Logical query File transfer requests

PPDG meeting, July 2000 Current PPDG test setup Query Estimator (QE) Hierarchical Resource Manager (HRM) U of Wisc Stage file (PFTP) File Catalog (FC) Bit- Sliced index Disk Cache file caching Client Sget, stage, sstatus, sabort Purge file Set of files LBNL Disk Cache Logical Query CORBA Storage Resource Broker (SRB) Note: there is no request manager in this figure. the Client is acting as its own the Request Manager

PPDG meeting, July 2000 SRB-HRM interaction Hierarchical Resource Manager (HRM) esrb.server Wisc Client 2 Wisc Client 3 fork IPC SRB Wisc Client 1 SRB child fork esrb.client IPC SRB child SRB esrb.client CORBA

PPDG meeting, July 2000 HRM esrb.server SRB esrb.client Write to FIFO: /tmp/lblsrb.512 fork FID: FID Call stage It has a table to store FID and requested file Name FID File Name path/run97 path/run100 When stage is done Write to FIFO: /tmp/lblsrb.512

PPDG meeting, July 2000 Status Functionality that was tested —request a file transfer - Sget —A single Client Functionality that exists but was not tested —Sget, Sstage, Sstatus —multiple Clients concurrent requests —use of replica catalog developed by SDSC Functionality that does not exist —A request manager that uses the replica catalog —A request manager that accesses multiple sites —A request manager that makes informed choices —A disk resource manager

PPDG meeting, July 2000 Optional Slides The following slides shows how HRM implements the functionality it provides They will be shown only if time permits

PPDG meeting, July 2000 Queuing File Transfers Number of PFTPs to HPSS are limited —limit set by a parameter - No_PFTP —parameter can be changed dynamically HRM is multi-threaded —issues and monitors multiple PFTPs in parallel All requests beyond PFTP limit are queued File Catalog used to provide for each file —HPSS path/file_name —Disk cache path/file_name —File size —tape ID

PPDG meeting, July 2000 File Queue Management Goal — minimize tape mounts — still respect the order of requests — do not postpone unpopular tapes forever File clustering parameter - FCP —If the file at top of queue is in Tape i and FCP > 1 (e.g. 4) then up to 4 files from Tape i will be selected to be transferred next —then, go back to file at top of queue Parameter can be set dynamically F 1 (T i ) F 3 (T i ) F 2 (T i ) F 4 (T i ) Order of file service

PPDG meeting, July 2000 File Caching Order for different File Clustering Parameters File Clustering Parameter = 1 File Clustering Parameter = 10

PPDG meeting, July 2000 Transfer Rate (Tr) Estimates Need Tr to estimate total time of a query Tr is average over recent file transfers from the time PFTP request is made to the time transfer completes. This includes: —mount time, seek time, read to HPSS Raid, transfer to local cache over network For dynamic network speed estimate —check total bytes for all file being transferred over small intervals (e.g. 15 sec) —calculate moving average over n intervals (e.g. 10 intervals)

PPDG meeting, July 2000 Query Estimate Given: transfer rate Tr. Given a query for which: —X files are in cache —Y files are in the queue —Z files are not scheduled yet Let s(file_set) be the total byte size of all files in file_set If Z = 0, then —QuEst = s(Y)/Tr If Z = 0, then —QuEst = (s(T)+ q.s(Z))/Tr where q is the number of active queries F 1 (Y) F 3 (Y) F 2 (Y) F 4 (Y) Y T

PPDG meeting, July 2000 Reason for q.s(Z) 20 Queries of length ~20 minutes launched 20 minutes apart 20 Queries of length ~20 minutes launched 5 minutes apart Estimate pretty close Estimate bad - request accumulate in queue

PPDG meeting, July 2000 Error Handling 5 generic errors —file not found return error to caller —limit PFTP reached can’t login re-queue request, try later (1-2 min) —HPSS error (I/O, device busy) remove part of file from cache, re-queue try n times (e.g. 3), then return error “transfer_failed” —HPSS down re-queue request, try repeatedly till successful respond to File_status request with “HPSS_down”

PPDG meeting, July 2000 File Tracking (1) HPSS down HPSS down

PPDG meeting, July 2000 File Tracking (2)

PPDG meeting, July 2000 Dynamic Display of Various Measurements