The GSI Mass Storage System TAB GridKa, FZ Karlsruhe Sep. 4, 2002 Horst Göringer, GSI Darmstadt

Slides:



Advertisements
Similar presentations
Categories of I/O Devices
Advertisements

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.
XenData SX-520 LTO Archive Servers A series of archive servers based on IT standards, designed for the demanding requirements of the media and entertainment.
Virtualisation From the Bottom Up From storage to application.
MUNIS Platform Migration Project WELCOME. Agenda Introductions Tyler Cloud Overview Munis New Features Questions.
Introduction to DBA.
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
The GSI Mass Storage for Experiment Data DVEE-Palaver GSI Darmstadt Feb. 15, 2005 Horst Göringer, GSI Darmstadt
Applying Data Grids to Support Distributed Data Management Storage Resource Broker Reagan W. Moore Ian Fisk Bing Zhu University of California, San Diego.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
1 I/O Management in Representative Operating Systems.
Operating Systems.
9 Copyright © Oracle Corporation, All rights reserved. Oracle Recovery Manager Overview and Configuration.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
Backup Rationalisation Reorganisation of the CERN Computer Centre Backups David Asbury IT/DS Friday 6 December 2002.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
1 Objectives Discuss the Windows Printer Model and how it is implemented in Windows Server 2008 Install the Print Services components of Windows Server.
Backup & Recovery 1.
© 2011 IBM Corporation 11 April 2011 IDS Architecture.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
1 Input/Output. 2 Principles of I/O Hardware Some typical device, network, and data base rates.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
Module 13: Configuring Availability of Network Resources and Content.
Chapter Oracle Server An Oracle Server consists of an Oracle database (stored data, control and log files.) The Server will support SQL to define.
CASPUR Site Report Andrei Maslennikov Sector Leader - Systems Catania, April 2001.
Maintaining File Services. Shadow Copies of Shared Folders Automatically retains copies of files on a server from specific points in time Prevents administrators.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
| nectar.org.au NECTAR TRAINING Module 9 Backing up & Packing up.
Chapter © 2006 The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/ Irwin Chapter 7 IT INFRASTRUCTURES Business-Driven Technologies 7.
MediaGrid Processing Framework 2009 February 19 Jason Danielson.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Hierarchical storage management
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
1 Week #10Business Continuity Backing Up Data Configuring Shadow Copies Providing Server and Service Availability.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
CHEP 2000: 7-11 February, 2000 I. SfiligoiData Handling in KLOE 1 CHEP 2000 Data Handling in KLOE I.Sfiligoi INFN LNF, Frascati, Italy.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
Process Architecture Process Architecture - A portion of a program that can run independently of and concurrently with other portions of the program. Some.
MISSION CRITICAL COMPUTING Siebel Database Considerations.
| nectar.org.au NECTAR TRAINING Module 9 Backing up & Packing up.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Unit 1: IBM Tivoli Storage Manager 5.1 Overview. 2 Objectives Upon the completion of this unit, you will be able to: Identify the purpose of IBM Tivoli.
© 2012 IBM Corporation IBM Linear Tape File System (LTFS) Overview and Demo.
An Introduction to GPFS
CommVault Architecture
CTA: CERN Tape Archive Rationale, Architecture and Status
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Compute and Storage For the Farm at Jlab
XenData SX-10 LTO Archive Appliance
CASTOR: possible evolution into the LHC era
Databases and DBMSs Todd S. Bacastow January 2005.
Storage Area Networks The Basics.
Network Attached Storage Overview
The INFN Tier-1 Storage Implementation
CTA: CERN Tape Archive Overview and architecture
Research Data Archive - technology
OffLine Physics Computing
XenData SX-550 LTO Archive Servers
Data, Databases, and DBMSs
Support for ”interactive batch”
CS703 - Advanced Operating Systems
CASTOR: CERN’s data management system
Backup Monitoring – EMC NetWorker
Backup Monitoring – EMC NetWorker
IBM Tivoli Storage Manager
STATEL an easy way to transfer data
Presentation transcript:

The GSI Mass Storage System TAB GridKa, FZ Karlsruhe Sep. 4, 2002 Horst Göringer, GSI Darmstadt

FZ Karlsruhe Sep. 4, Mass GSI: Overview  recent history  current (old) system  functionality  structure  usage  new system  the requirements  structure  status  Outlook

FZ Karlsruhe Sep. 4, Mass GSI: History till 1995: ATL Memorex 5400, capacity 1000 Gbyte IBM 3480 cartridges IBM HSM (MVS) 1995: IBM 3494 ATL, ADSM (AIX) ADSM interface not acceptable: - no cross-platform support for clients (AIX, VMS) - ADSM "node-based" - no guaranteed availability for staged files - ADSM commands not so easy to use => use ADSM API for GSI mass storage system (1996) 2000: broad discussions on future system 2001: decision to enhance GSI mass storage system

FZ Karlsruhe Sep. 4, Mass GSI: Server current server hardware:  IBM RS6000 H50, 2 processors, 1 GB memory  Gigabit Ethernet (max rate ~21 Mbyte/s)  IBM 3494 Automatic Tape Library  8 IBM 3590E tapes (capacity GByte uncompressed)  current capacity ATL 60 TByte (currently used: 27 TB exp. data, 12 TB backup)  ~350 GByte internal staging disks current server software:  AIX V4.3  Tivoli Storage Manager Server V4.1.3

FZ Karlsruhe Sep. 4, Mass GSI: Logical Structure

FZ Karlsruhe Sep. 4, Mass GSI: Functionality 1.command interface  archive  retrieve  query  stage  delete  pool_query  ws_query 2. RFIO API  functions available for open, close, read, seek,...  RFIO client accesses GSI mass storage  write functions: till end 2002

FZ Karlsruhe Sep. 4, Mass GSI: Functionality identical client interface on all GSI platforms (Linux, AIX, VMS) unique name space security policy client tape support (ANSI Label) server log files for - error reports - statistical analysis GSI software: C with sockets integrated in Alien

FZ Karlsruhe Sep. 4, Mass GSI: Stage Pool Manager  administers several Stage Pools with different attributes:  file life time  max space  user access  currently active pools:  RetrievePool: no guaranteed life time  StagePool: min life time guaranteed  future pools:  ArchivePool  pools dedicated to user groups

FZ Karlsruhe Sep. 4, Mass GSI: Stage Pool Manager  administers an additional stage meta data DB  locks each access to a pool  handles disk clean requests from different sources:  from process serving a user  from watch demon  initiates and controls disk clean processes

FZ Karlsruhe Sep. 4,

FZ Karlsruhe Sep. 4, Mass GSI: Upgrade Requirements scalable system needed (data capacity and max data rate) 1. higher bandwith => several data movers oeach with access to each tape device and robot oeach with own disk pools 2. one master administering the complete meta data DB 3. hardware independency This means:  fully parallel data streams  separation of control flow and data flow

FZ Karlsruhe Sep. 4, Mass GSI: Upgrade Requirements enabling technologies:  Storage Area Network  Tivoli Storage Manager (successor of ADSM)

FZ Karlsruhe Sep. 4, Mass GSI: New Structure

FZ Karlsruhe Sep. 4, Mass GSI: New Hardware new hardware:  tape robot StorageTek L700 (max 68 TByte)  8 IBM 3850 Ultrium LTO tape drives (capacity 100 GByte uncompressed)  2 TSM server (Intel PC, fail-safe Windows 2000 cluster)  4 (8) data mover (Intel PC, Windows 2000)  SAN components Brokat switch 4100 (16 ports, each 1 Mbit/s) purpose:  verification of new concept  hardware test: SAN, ATL, tape drives, tape volumes  later: new backup system for user data

FZ Karlsruhe Sep. 4, Mass GSI: Status hardware, TSM/Storage Agent: seems to work (tests still running) new GSI software (for Unix) nearly ready command client RFIO client (read only) server package (master and slaves on data movers) stage pool manager (master and slaves on data movers) in Oct 2002: to be used for production with current AIX server

FZ Karlsruhe Sep. 4, Mass GSI: Current Plans  in 2003: DAQ connection to mass storage  n event builders will write in parallel via RFIO to dedicated archive disk pools  enhanced performance and stability requirements  in 2003/2004: new ATL (several 100 TByte) to fulfill current requirements for next years

FZ Karlsruhe Sep. 4, Mass GSI: Outlook  the yearly increment of experiment data grows rapidly:  an order of magnitude next years  after 2006 Alice experiment running  "Future Project" of GSI => the mass storage system must be scalable in both, storage capacity, and data rates  the system must be flexible to follow the development of new hardware Our new concept fulfills these requirements!  TSM is a powerful storage manager satisfying our needs now and also in the next future  high flexibility with GSI made user interface

Mass GSI Appendix More Details

FZ Karlsruhe Sep. 4, Mass GSI: DAQ Connection

FZ Karlsruhe Sep. 4, Mass GSI : the new System not only upgrade - entrance into new hardware and new platform!  new server platform Windows 2000  new tape robot  new tape drives and media  new network more work necessary due to  missing practice  unknown problems  lower quality of tape drives and media  presumably more operation failures => costs reduced by cheaper components, but more manpower necessary (in development and operation) however, we have many options for the future!

FZ Karlsruhe Sep. 4, Mass GSI : the new System SW enhancements and adaptions (cont‘d):  adaption adsmcli server to new concept => tsmcli  division of functionality into several processes  code restructuring and adaptions  communication between processes  data mover selection (load balancing)  enhancement disk pool manager  subpools on each data mover => n slave disk pool managers  communication master - slaves  enhancement metadata database  subpool selection  DAQ pool handling

FZ Karlsruhe Sep. 4, Mass GSI : the new System Potential risks:  new server platform Windows 2000  new tape robot  new tape drives and media  new network more work necessary due to  missing practice  unknown problems  lower quality of tape drives and media

FZ Karlsruhe Sep. 4, Current Status: File Representation  file archived is defined by o archive name o path name (independent from local file system) o file name (identical with name in local file system)  user access handled by access table for all supported client platforms  files already archived are not overwritten o except explicitly required  local files are not overwritten (retrieve)

FZ Karlsruhe Sep. 4, Current Status: Local Tape Handling  support of standard ANSI Label tapes on client side  tape volumes portable between client platforms (AIX, Linux, VMS)  enhanced error handling: ocorrupt files do not affect others when handling a file list omissing EOF mark is handled  user friendly: archive complete tape volume invoking one command

FZ Karlsruhe Sep. 4, Current Status: why Disk Pools disk pools help to avoid 1. tape mount and load times 2. concurrent access to same tape volume 3. blocking of fast tape drives in robot this is useful if  files are needed several times in a short time  the network connection of clients is slow  large working sets are retrieved file after file  large working sets are accessed in parallel from a compute farm

FZ Karlsruhe Sep. 4,

FZ Karlsruhe Sep. 4, Disk Pool Manager: the current Pools  RetrievePool: o files stored via adsmcli retrieve - inhibit with option stage=no o files stored when read via API o no guaranteed life time  StagePool: o files stored via adsmcli stage o min life time guaranteed (3 days currently)  current space: o hardware shared o overall: 350 GByte o StagePool: 100 GByte max o the RetrievePool uses space unused by the StagePool (`lent space`)

FZ Karlsruhe Sep. 4, API Client for Mass Storage API client: there are functions available e.g.  to open/close files in mass storage  to read/write buffers in remote files  to shift the file pointer in remote files => data stream: analysis program - mass storage fully controlled by user useful if  only selective access to (small) parts of file required oparts of ROOT tree ontuples  local disk space unsufficient requirement for GSI API client:  compatible with CERN/HEP RFIO package RFIO interface available in CERN applications

FZ Karlsruhe Sep. 4, API Client: Logical Structure

FZ Karlsruhe Sep. 4, API Client: RFIO at GSI RFIO functions developed: needed for ROOT: rfio_open, rfio_read, rfio_close, rfio_lseek additionally (e.g. for analysis programs): rfio_fopen, rfio_fread, rfio_fclose file name: file representation in GSI mass storage system currently available at GSI: –enhanced adsmcli server already in production –sample C program using rfio_f... (read) on Linux: /GSI/staging/rfio –ROOT with RFIO client (read) –GO4 viewer with RFIO client (read) in future: write functionality (rfio_write, rfio_fwrite)  presumably more operation failures (also data movers) => costs reduced by cheaper components more manpower necessary (development and operation)

FZ Karlsruhe Sep. 4, API Client: ROOT with RFIO ROOT at GSI since version with RFIO API For RFIO usage in ROOT:  load shared library libRFIO.so in your ROOT session  for file open: use class TRFIOFile instead of TFile  prefix the file representation in the mass storage system with `rfio:` in GO4 viewer: no prefix to the file representation needed

FZ Karlsruhe Sep. 4,

FZ Karlsruhe Sep. 4,

FZ Karlsruhe Sep. 4, Mass GSI : The current Bottlenecks data capacity:  in April 2001: tape robot nearly completely filled (30 TByte uncompressed)  since April 27, 2001: new tape drives IBM 3590E o write with double density: > 20 GByte/volume o copy all volumes => ~ 30 TByte free capacity  current requirements GSI (TByte): experiment backup accumulated 2001: : :  additionally: multiple instances of experiment data!