Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop, 29.11.07.

Slides:



Advertisements
Similar presentations
Author - Title- Date - n° 1 GDMP The European DataGrid Project Team
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
The LEGO Train Framework
GSIAF "CAF" experience at GSI Kilian Schwarz. GSIAF Present status Present status installation and configuration installation and configuration usage.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Statistics of CAF usage, Interaction with the GRID Marco MEONI CERN - Offline Week –
New CERN CAF facility: parameters, usage statistics, user support Marco MEONI Jan Fiete GROSSE-OETRINGHAUS CERN - Offline Week –
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Staging to CAF + User groups + fairshare Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE Offline week,
Microsoft ® Official Course Module 12 Monitoring, Managing, and Recovering AD DS.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
AliEn uses bbFTP for the file transfers. Every FTD runs a server, and all the others FTD can connect and authenticate to it using certificates. bbFTP implements.
OSG Public Storage and iRODS
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
CERN IT Department CH-1211 Genève 23 Switzerland t EIS section review of recent activities Harry Renshall Andrea Sciabà IT-GS group meeting.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Status of the production and news about Nagios ALICE TF Meeting 22/07/2010.
Module 7: Resolving NetBIOS Names by Using Windows Internet Name Service (WINS)
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
PROOF Cluster Management in ALICE Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
AgentsAnd Daemons Automating Data Quality Monitoring Operations Agents And Daemons Automating Data Quality Monitoring Operations Since 2009 when the LHC.
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
Status of PDC’07 and user analysis issues (from admin point of view) L. Betev August 28, 2007.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
Module 8 : Configuration II Jong S. Bok
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
EGI-InSPIRE EGI-InSPIRE RI DDM solutions for disk space resource optimization Fernando H. Barreiro Megino (CERN-IT Experiment Support)
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS XROOTD news New release New features.
10 May 2001WP6 Testbed Meeting1 WP5 - Mass Storage Management Jean-Philippe Baud PDP/IT/CERN.
Data Management The European DataGrid Project Team
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
Status of the Shuttle Framework Alberto Colla Jan Fiete Grosse-Oetringhaus ALICE Offline Week October 2006.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks The LCG interface Stefano BAGNASCO INFN Torino.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
03/09/2007http://pcalimonitor.cern.ch/1 Monitoring in ALICE Costin Grigoras 03/09/2007 WLCG Meeting, CHEP.
ALICE experiences with CASTOR2 Latchezar Betev ALICE.
CASTOR Operations Face to Face 2006 Miguel Coelho dos Santos
CERN - IT Department CH-1211 Genève 23 Switzerland CASTOR F2F Monitoring at CERN Miguel Coelho dos Santos.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
Good user practices + Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CUF,
CASTOR new stager proposal CASTOR users’ meeting 24/06/2003 The CASTOR team.
VO Box discussion ATLAS NIKHEF January, 2006 Miguel Branco -
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
Cluster Status & Plans Cluster Status & Plans —— Gang Qin Jan
The ALICE Analysis -- News from the battlefield Federico Carminati for the ALICE Computing Project CHEP 2010 – Taiwan.
Federating Data in the ALICE Experiment
Experience of PROOF cluster Installation and operation
Real Time Fake Analysis at PIC
OpenPBS – Distributed Workload Management System
Report PROOF session ALICE Offline FAIR Grid Workshop #1
Status of the CERN Analysis Facility
GSIAF & Anar Manafov, Victor Penso, Carsten Preuss, and Kilian Schwarz, GSI Darmstadt, ALICE Offline week, v. 0.8.
Offline shifter training tutorial
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
GSIAF "CAF" experience at GSI
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
CTA: CERN Tape Archive Overview and architecture
Support for ”interactive batch”
Overview Multimedia: The Role of WINS in the Network Infrastructure
Presentation transcript:

Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus2 Disk Buffer Tier-1 data export CAF computing cluster Proof worker local disk Proof master, xrootd redirector Proof worker local disk Proof worker local disk... Staging AliEn SE CASTOR MSS Tape CAF Schema

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus3 Staging Files are produced in ALICE's PDC and stored in AliEn SEs (for CERN: CASTOR) Step 1 (first months): Manual –Files copied by a shell script to redirector that balances between disk servers –To allow user staging the nodes were open for writing –Complicated for users, no control over quotas, difficult garbage collection Step 2 (until mid 2007): Semi-automatic –Staging script plugged into xrootd Prepare request with stage flag or open request to a file triggered staging –User gets list of files from the AliEn FC and triggers staging for all files –Convenient for users, no quotas, difficult garbage collection

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus4 Staging (2) Step 3 (now): Automatic –Staging script plugged into olbd –Implementation of PROOF datasets (by ALICE) –Staging daemon that runs on the cluster –Transparent migration from AliEn collection to PROOF datasets –Convenient for users, quota-enabled, garbage collection 3 TB ( files) staged to the system

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus5 Introduction of PROOF datasets A dataset represents a list of files (e.g. physics run X) –Correspondence between AliEn collection and PROOF dataset Users register datasets –The files contained in a dataset are automatically staged from AliEn (and kept available) –Datasets are used for processing with PROOF Contain all relevant information to start processing (location of files, abstract description of content of files) File-level storing by underlying xrootd infrastructure Datasets are public for reading (you can use datasets from anybody!) There are common datasets (for data of common interest)

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus6 PROOF worker / xrootd disk server (many) PROOF master / xrootd redirector PROOF master Dataset registers dataset removes dataset uses dataset data manager daemon data manager daemon keeps dataset persistent by requesting staging updating file information touching files olbd/ xrootd olbd/ xrootd file stager stages files removes files that are not used (least recently used above threshold) selects disk server and forwards stage request WN disk … write delete read read, touch stage Dataset concept AliEn SE CASTOR MSS

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus7 Staging script Two directories configured in xrootd/olbd for staging –/alien –/castor Staging script given with olb.prep directive –Perl script that consists of 3 threads –Front-End: Registers stage request –Back-End Checks access privileges Triggers migration from tape (CASTOR, AliEn) Copies files, notifies xrootd –Garbage collector: Cleans up following policy file with low/high watermarks (least recently used above threshold)

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus8 Data manager daemon Keeps content of datasets persistent on disk Regularly loops over all datasets Sends staging requests for new files Extracts meta data from recently staged files Verifies that all files are still available on the cluster (by touch, prevents garbage collection) –Speed: 100 files / s

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus9 PROOF master Registering, removal of datasets –Checks quota upon registration (group level quotas) Display datasets, quotas Use datasets –Meta data contained in dataset allows to skip lookup and validation step

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus10 Datasets in Practice Create DS from AliEn collection –collection = TGrid::OpenCollection(lfn) –ds = collection->GetFileCollection() Upload to PROOF cluster –gProof->RegisterDataSet("myDS", ds) Check status: gProof->ShowDataSet("myDS") Use it: gProof->Process("myDS", "mySelector.cxx+") (not completely implemented)

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus11 Dataset in Practice (2) List available datasets: gProof->ShowDataSets() You always see common datasets and datasets of your group This method was used to stage 3 M events of PDC07

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus12 Monitoring of datasets Data set usage per group Number of files per host

Dynamic staging to a CAF cluster - Jan Fiete Grosse-Oetringhaus13 Status Staging script implemented and in use since a year Daemon in place and running since 1-2 months Dataset handling in PROOF implemented (by ALICE), in ROOT SVN, but in an own development branch Processing of datasets in prototype stage (to be implemented by PROOF team)