CASTOR new stager proposal CASTOR users’ meeting 24/06/2003 The CASTOR team.

Slides:



Advertisements
Similar presentations
1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Advertisements

Study of Hurricane and Tornado Operating Systems By Shubhanan Bakre.
Resource Management with YARN: YARN Past, Present and Future
16/9/2004Features of the new CASTOR1 Alice offline week, 16/9/2004 Olof Bärring, CERN.
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
16/4/2004Storage Resource Sharing with CASTOR1 Olof Barring, Benjamin Couturier, Jean-Damien Durand, Emil Knezo, Sebastien Ponce (CERN) Vitali Motyakov.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Resource Management and Accounting Working Group Working Group Scope and Components Progress made Current issues being worked Next steps Discussions involving.
A Lightweight Platform for Integration of Resource Limited Devices into Pervasive Grids Stavros Isaiadis and Vladimir Getov University of Westminster
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting June 13-14, 2002.
Building Advanced Storage Environment Cheng Yaodong Computing Center, IHEP December 2002.
Frascati, October 9th, Accounting in DataGrid Initial Architecture Albert Werbrouck Frascati, October 9, 2001.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
Olof Bärring – WP4 summary- 4/9/ n° 1 Partner Logo WP4 report Plans for testbed 2
May PEM status report. O.Bärring 1 PEM status report Large-Scale Cluster Computing Workshop FNAL, May Olof Bärring, CERN.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Functional description Detailed view of the system Status and features Castor Readiness Review – June 2006 Giuseppe Lo Presti, Olof Bärring CERN / IT.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Chapter 2 Processes and Threads Introduction 2.2 Processes A Process is the execution of a Program More specifically… – A process is a program.
Integrated Monitoring Dashboard SEAP/BTIS/DB&I/Mar Integrated Monitoring Dashboard An Innovative Monitoring Approach Proposal by BTIS/DB&I.
New stager commands Details and anatomy CASTOR external operation meeting CERN - Geneva 14/06/2005 Sebastien Ponce, CERN-IT.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CASTOR status Presentation to LCG PEB 09/11/2004 Olof Bärring, CERN-IT.
1 LHCb File Transfer framework N. Brook, Ph. Charpentier, A.Tsaregorodtsev LCG Storage Management Workshop, 6 April 2005, CERN.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
G.Govi CERN/IT-DB 1 September 26, 2003 POOL Integration, Testing and Release Procedure Integration  Packages structure  External dependencies  Configuration.
Aneka Cloud ApplicationPlatform. Introduction Aneka consists of a scalable cloud middleware that can be deployed on top of heterogeneous computing resources.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS New tape server software Status and plans CASTOR face-to-face.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Distributed Logging Facility Castor External Operation Workshop, CERN, November 14th 2006 Dennis Waldron CERN / IT.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
Douglas Thain, John Bent Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, Miron Livny Computer Sciences Department, UW-Madison Gathering at the Well: Creating.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
CERN - IT Department CH-1211 Genève 23 Switzerland CASTOR F2F Monitoring at CERN Miguel Coelho dos Santos.
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
CASTOR in SC Operational aspects Vladimír Bahyl CERN IT-FIO 3 2.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Cluman: Advanced Cluster Management for Large-scale Infrastructures.
The Holmes Platform and Applications
CASTOR: possible evolution into the LHC era
Databases and DBMSs Todd S. Bacastow January 2005.
CASTOR Giuseppe Lo Presti on behalf of the CASTOR dev team
(on behalf of the POOL team)
High Availability Linux (HA Linux)
Integrating HA Legacy Products into OpenSAF based system
Giuseppe Lo Re Workshop Storage INFN 20/03/2006 – CNAF (Bologna)
Dirk Düllmann CERN Openlab storage workshop 17th March 2003
Ákos Frohner EGEE'08 September 2008
The Scheduling Strategy and Experience of IHEP HTCondor Cluster
Chapter 2: Database System Concepts and Architecture
Data, Databases, and DBMSs
Support for ”interactive batch”
Database Environment Transparencies
CASTOR: CERN’s data management system
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

CASTOR new stager proposal CASTOR users’ meeting 24/06/2003 The CASTOR team

24/06/2003 CASTOR new stager proposal, 2 Outline The vision... and a caveat Some problems with today’s system Proposal –Ideas how to get around those problems –Architecture –Request registration and scheduling –Catalogues –Disk server access and physical file ownership –Some interesting features Project planning and progress monitoring Conclusions

24/06/2003 CASTOR new stager proposal, 3 Vision... With clusters of 100s of disk and tape servers, the automated storage management faces more and more the same problems as CPU clusters management –(Storage) Resource management –(Storage) Resource sharing –(Storage) Request scheduling –Configuration –Monitoring The stager is the main gateway to all resources managed by CASTOR Vision: Storage Resource Sharing Facility

24/06/2003 CASTOR new stager proposal, and a caveat The vision is to provide a scalable Storage Resource Sharing Facility –The hope is to achieve similar efficiency for the storage resource utilization as LSF provides for the CPU resources today However: nothing in the proposed design enforces a single shared stager instance –Today’s configurations with some 40 independent stagers is still OK

24/06/2003 CASTOR new stager proposal, 5 Some problems today’s stager Lot of code for supporting direct tape access No true request scheduling –Throttling, load-balancing –Fair-share Resource sharing not supported –Stagers are either dedicated or public –Dedicated resources  Some disk servers are 100% full/loaded while others are idle –Public resources  No control of who gets how much of the resources. Invites for abuse Operational issues –No unique request identifiers –Problem tracing difficult stagein –V P01234 –v EK4432 –q u –f MYHIGGSTODAY \ -g 994BR5 –b 8000 –F FB –L 80 –C ebcdic,block –E skip

24/06/2003 CASTOR new stager proposal, 6 PROPOSAL Proposal Ideas for the new stager Pluggable framework rather than total solution –True request scheduling: delegate the scheduling to a pluggable black-box scheduler. Possibly using third party schedulers, e.g. Maui or LSF –Policy attributes: externalize policy engines governing the resource matchmaking. Start with today’s policies for file system selection, GC, migration,.... Could move toward full-fledged policy languages, e.g. implemented using “GUILE” Restricted access to storage resources to achieve predictable load –No random rfiod eating up the resources behind the back of the scheduling system Disk server autonomy as far as possible –In charge of local resources: file system selection and execution of garbage collection –Loosing a server should not affect the rest of the system

24/06/2003 CASTOR new stager proposal, 7 PROPOSAL Physics application RFIO API stage API Request scheduler Request queue Common RTCOPY client: rtcpclientd mstaged VDQM VMGR Cns Global catalogue Scheduling policies Local catalogue Local policies sstaged rfiod Get physical path rtcpd Get disk server Start tape request rfiod Access /castor/… Request /castor/… Local Request scheduler RequestHandler New module Existing, modified External Existing module

24/06/2003 CASTOR new stager proposal, 8 PROPOSAL Proposal Request scheduling (1) A “master stager” (mstaged) receives all CASTOR file access requests –Authenticate client and register the request –Queue the request –The request registration is independent of the scheduling. It has to be designed to cope with high request load peaks Pluggable scheduler manages the queue and applies configured policies –E.g. requests from gid=1307 should only run on atlas001d,...

24/06/2003 CASTOR new stager proposal, 9 PROPOSAL Proposal Request handling & scheduling RequestRegister Fabric Authentication service e.g. Kerberos-V server Read: /castor/cern.ch/user/c/castor/TastyTreesDN=castor Typical file request Thread pool Authenticate “castor” Request repository (Oracle, MySQL) Scheduler Scheduling Policies user “castor” has priority Dispatcher Store request Run request on pub003d Get Jobs Disk server load Catalogue File staged? Request registration: Must keep up with high request rate peaks Request scheduling: Must keep up with average request rates

24/06/2003 CASTOR new stager proposal, 10 PROPOSAL Proposal Request scheduling (2) A “slave stager” (sstaged) runs on each disk server –Executes and controls all requests scheduled to it by the mstaged –Takes care of local resource scheduling such as file system selection and execution of garbage collector The sstaged also gathers relevant local load information for the central scheduler

24/06/2003 CASTOR new stager proposal, 11 PROPOSAL Proposal Catalogues Request catalogues –Central Repository of all running requests + request history Predictable load  facilitate load balancing Usage accounting from request history Fair-share File catalogues –Central CASTOR file  disk server mapping allows for finding files –Local CASTOR file  physical filename catalogue on the disk servers

24/06/2003 CASTOR new stager proposal, 12 PROPOSAL Proposal Disk server access Today a user can access files on disk servers either by –The CASTOR file name /castor/cern.ch/... –The physical file name /shift/lhcb003d/... With the new stager we restrict –To only allow for access by CASTOR file name –All physical files are owned by a generic account (stage,st) and their paths are hidden from direct RFIO access WHY????

24/06/2003 CASTOR new stager proposal, 13 PROPOSAL Proposal Disk server access Avoid two databases for file permissions & ownership –CASTOR name server –File system holding physical file Facilitate migration/recall of user files –Files with different owners are normally grouped together on tapes owned by a generic account (stage,st) –Would like to avoid setuid/setgid for every file Avoid backdoors: all disk server access must be scheduled An useful analogy: forbid interactive login access to the batch nodes in a LSF cluster

24/06/2003 CASTOR new stager proposal, 14 PROPOSAL Proposal Disk server access File: /castor/cern.ch/user/c/castor/RottenTrees Owner: castor,c3 File: /shift/pub002d/data05/c3/stage/castor/RottenTress Owner: stage,st CASTOR name server sstaged managed disk server All scheduled access result in an instance of a rfiod running on behalf of the user under generic account (stage,st) on the disk server rfiod authenticates the user and checks that the request has been scheduled. Unscheduled requests are rejected rfiod on disk server only allows for access to /castor files. Access with physical path is rejected

24/06/2003 CASTOR new stager proposal, 15 PROPOSAL Proposal Some interesting features Modifications to the tape mover allows for adding files to running tape requests Migration (and recall) controlled by a new central component called rtcopyclientd –Initiates the tape requests –Schedules the file copies just-in-time when the tape is positioned Dynamically expanding migration streams Better load-balancing is possible since the files copies are scheduled according to the load Allow for seeks in RFIO v3 (streaming) mode

24/06/2003 CASTOR new stager proposal, 16 Project planning and monitoring Detailed plan in proposal document Three milestones: –October -03: Demonstrate concept of pluggable scheduler and high rate request handling –February -04: Integrated prototype of the whole system –April -04: Production system ready for deployment Progress monitoring –Aim to use Project/Task manager provided by LCG Savannah portal ( –Progress reviews at each milestone? are the experiments interested in providing efforts for helping with review?

24/06/2003 CASTOR new stager proposal, 17 Conclusions The proposal aims for –A pluggable framework for intelligent and policy controlled file access scheduling –Evolvable storage resource sharing facility framework rather than a total solution –File access request running/control and local resource allocation delegated to disk servers Questions, remarks, proposals?