Proposal for dCache based Analysis Disk Pool for CDF presented by Doug Benjamin Duke University on behalf of the CDF Offline Group.

Slides:



Advertisements
Similar presentations
Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Advertisements

Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
1 Storage Today Victor Hatridge – CIO Nashville Electric Service (615)
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Extensible Scalable Monitoring for Clusters of Computers Eric Anderson U.C. Berkeley Summer 1997 NOW Retreat.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Minerva Infrastructure Meeting – October 04, 2011.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Tier 3g Infrastructure Doug Benjamin Duke University.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
LAN / WAN Business Proposal. What is a LAN or WAN? A LAN is a Local Area Network it usually connects all computers in one building or several building.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
LcgCAF:CDF submission portal to LCG Federica Fanzago for CDF-Italian Computing Group Gabriele Compostella, Francesco Delli Paoli, Donatella Lucchesi, Daniel.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
©2006 Merge eMed. All Rights Reserved. Energize Your Workflow 2006 User Group Meeting May 7-9, 2006 Disaster Recovery Michael Leonard.
Tier 3(g) Cluster Design and Recommendations Doug Benjamin Duke University.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
08/30/05GDM Project Presentation Lower Storage Summary of activity on 8/30/2005.
Database Architectures Database System Architectures Considerations – Data storage: Where do the data and DBMS reside? – Processing: Where.
Computing at CDF ➢ Introduction ➢ Computing requirements ➢ Central Analysis Farm ➢ Conclusions Frank Wurthwein MIT/FNAL-CD for the CDF Collaboration.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Marco Cattaneo - DTF - 28th February 2001 File sharing requirements of the physics community  Background  General requirements  Visitors  Laptops 
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
May Donatella Lucchesi 1 CDF Status of Computing Donatella Lucchesi INFN and University of Padova.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Compute and Storage For the Farm at Jlab
Scalable sync-and-share service with dCache
Artem Trunov and EKP team EPK – Uni Karlsruhe
LQCD Computing Operations
Presentation transcript:

Proposal for dCache based Analysis Disk Pool for CDF presented by Doug Benjamin Duke University on behalf of the CDF Offline Group

10-October-2006CDF Analysis Disk Pool - D Benjamin2 Outline The Problem & Proposed Solution Deployment Process Collaborative and phased approach with feedback from users and experts at each stage

10-October-2006CDF Analysis Disk Pool - D Benjamin3 How CDF does data analysis Collision Data Flow Raw data written to tape Measure the final calibrations Full and final reconstruction on production farm or CDF user analysis cluster (CAF or FermiGrid) Input read in from tape and output written to tape Centralized ntuple production (most CDF physicists access the data through ntuples) Collaboration wide ntuples produced on production farm or CDF CAF Physics groups further reduce the centralized ntuples or production files (strip off types of events) for further study User analysis of ntuples and production data run on CAF or FermiGrid Final Analysis on user desktops or institutional machines

10-October-2006CDF Analysis Disk Pool - D Benjamin4 How CDF does data analysis (2) Simulation Data Flow (Monte Carlo) Monte Carlo generation and reconstruction on remote computing Offsite dCaf’s NAMCAF (OSG sites - including Fermi Lab) LCGCAF (Europe) Files return to FNAL for storage on archival media (tape) Small files (< 1 GB in size) must be merged together MC files written to tape Centralized ntuple production run on MC Samples Some physics groups produce ntuples at time of generation/reconstruction Other groups create ntuples after the MC samples are on tape MC ntuple files are merged together before being written to tape Like Collison data - Physics groups further reduce the centralized ntuples or \MC files (strip off types of events) for further study User analysis of MC ntuples and data performed on CAF or FermiGrid Final Analysis on user desktops or institutional machines

10-October-2006CDF Analysis Disk Pool - D Benjamin5 Physics groups Project Disk space needs Physics groups need project space 5 Groups: B, Top, Exotic, QCD, EWK Several use cases : secondary or tertiary datasets creation and use User desktop Analysis results Tape backed dCache CAF nodes Stripped files Analysis Disk pool User analysis CAF nodes Creation and use of secondary Data/MC/Ntuple datasets

10-October-2006CDF Analysis Disk Pool - D Benjamin6 Small Analysis Group disk space needs Small Analysis Group project space (a few people working on an specific analysis) Analysis specific MC - Special study samples User desktop Analysis results CAF nodes User analysis Off-site Computers Concatenated files Analysis Disk pool CAF nodes Specialty Monte Carlo Samples Small files

10-October-2006CDF Analysis Disk Pool - D Benjamin7 Offline Monte Carlo Group disk space needs Concatenation of MC data files and ntuple files produced offsite File size requirements in data handling system (> 1 GB) Maximum CPU time requirements of Monte Carlo jobs limit output file size Implies files must be concatenated before uploading into DH system Tape backed dCache SAM upload Concatenated files Analysis Disk pool CAF nodes Production Monte Carlo Concatenation Analysis Disk pool Off-site Computers Small files

10-October-2006CDF Analysis Disk Pool - D Benjamin8 Data to be stored Secondary and Tertiary datasets production data, MC data and ntuples Stripped from data held in data handling system Temporary data Data used during the concatenation process MC data used for small analyses that can be easily regenerated (Note - archival quality data will be stored in the data handling system)

10-October-2006CDF Analysis Disk Pool - D Benjamin9 Current situation CDF is using discrete independent fileservers. ~70 fileservers for project space (> 160 TB) Can not easily Centrally manage the space Space inefficiently used (stale data) Data migration requires access to individual nodes and users have to change all their scripts Individual files servers approach not scalable Users often overload one or two file servers No Load balancing

10-October-2006CDF Analysis Disk Pool - D Benjamin10 Physics groups space needs Polled physics groups to get space needs per fb -1 groups space per fb -1 Top15 TB QCD12 TB Exotics20 TB B group8 TB Electroweak10 TB MC production20 TB Total 85 TB/fb TB is sufficient for data taken to date TB should be sufficient through FY ‘09

10-October-2006CDF Analysis Disk Pool - D Benjamin11 Project Disk requirements Stability, simplicity and reduced operational effort Scaleable - (add space/users as needed) Centralized administration to monitor and maximize our limited resources Simple user code changes (avoid lose in productivity) Decouple file storage from compute nodes Network file access (URL based) Avoids strong coupling between CAF worker nodes and fileservers hosting data

10-October-2006CDF Analysis Disk Pool - D Benjamin12 Proposed Solution Replace the majority of the static disk space with dCache based pool (the analysis disk pool) Use it for large files for which dCache is known to work well Store small files, e.g. log files on other disk based system e.g. on nfs mounted disks visible from Interactive Login Pool nodes dCache analysis diskpool finished prototype testing phase 4 head nodes for the services 9 fileservers (~ 50 TB total) 5 oldest filesersevers (~ 4 years old) to be replaced after this review

10-October-2006CDF Analysis Disk Pool - D Benjamin13 Advantages Solution adopted/supported at Fermilab and within HEP Allows for unified support and expertise sharing groups using dCache (DESY, CMS, CDF…) Global space management more efficient use of disk resources more transparent maintainability Decoupling of the name and disk space Scalability User files spread across several servers Client software already used by CDF

10-October-2006CDF Analysis Disk Pool - D Benjamin14 Risks and Risk remediation In a centralized system a small group of users may inadvertently affect all users User education reduces the risk Have not had any user induced outages to date Added software layer on top of storage hardware Centralized resources stability Production dCache system fairly stable (some problems with fileservers - solved by proper hardware drivers and retirement of the oldest nodes) File server stability can affect large number of users. Spread data across several servers Minimizes data inaccessibility due to one server down Balances the load

10-October-2006CDF Analysis Disk Pool - D Benjamin15 Risks and Risk remediation(2) What if some component of dCache changes underneath us (like namespace change: PNFS to Chimera)? Not a problem Chimera has a filesystem core onto of a relational database. To the CDF end user he/she won’t see much of a difference What if dCache is replaced with technology Y? If a root TFile class for technology Y exists then minor change to CDF software (URL’s to access files across the network) If namespace looks like a filesystem then mininal change to CDF software

10-October-2006CDF Analysis Disk Pool - D Benjamin16 Risk Management Many fileservers to reduce the dependence on only a few parts. All data stored on system could be restored or regenerated with reasonable amount of effort. Monitor the system to make sure it stays within the stable limits Establish usage guidelines and limit exposures when possible (e.g. limit pnfs mounts) Use good quality hardware for head nodes (have a spare node capable of being put into service quickly)

10-October-2006CDF Analysis Disk Pool - D Benjamin17 Usage guidelines Minimum file size 1 GB for long lived files Small files to be concatenated are excluded Prohibit use of tar, unzip and similar operations on Disk Pool Prohibit compilation, linking, debugging and similar activities Restrict large scale use of ls on pnfs namespace as a general user option. Browsable listings of directories and files will be created at least once per day and visible on the web

10-October-2006CDF Analysis Disk Pool - D Benjamin18 Usage guidelines(2) Prohibit the storage of log files in Disk Pool Log files are to be stored on static disks in the CDF interactive login pool (ILP) Prohibit looping over multiple files at an unduly high rate (for example TChain in Root or other techiniques) File transaction rate should be limited to 5 Hz across the entire system Users agree to use the Disk Pool within the established techincal specifications

10-October-2006CDF Analysis Disk Pool - D Benjamin19 Production system Specifications Total data volume served per day - 17 TB/day Based on average current University File server system (150 MB/s) and prototype disk pool system (40 MB/s) Aggregate bandwidth MB/s Peak of University system MB/s and prototype disk pool - 80 MB/s Maximum supported transaction rate 5 Hz Expected transaction rate ~ 0.2 Hz Maximum number of concurrent jobs 700

10-October-2006CDF Analysis Disk Pool - D Benjamin20 Daily data volume (TB) by CDF dCache systems

10-October-2006CDF Analysis Disk Pool - D Benjamin21 Analysis Disk pool - # of CDF jobs/hour 7 days 31 days Total Active movers (jobs) - blue line < 300 jobs per hour Total data volume ~ 50 TB Peak read rate: < 120 MB/s Note - number of jobs/ transaction rate will scale with amount of CPU cycles availible (worker nodes) - (not data volume stored) - Expected CDF CPU growth 200% now to FY’09

10-October-2006CDF Analysis Disk Pool - D Benjamin22 Proposed Support Model Three tier approach: Day-to-day operations and trouble shooting by CDF power users & CDF offline operations managers Diagnosing difficult problems, evolving and reconfiguring system when needed by a group from a CDF institution (per a to-be established MOU) – also serving as a point of contact between the experiment and CD Expert level consultations within to-be agreed upon limits by CD dCache development team

10-October-2006CDF Analysis Disk Pool - D Benjamin23 Proposed Support Model (cont’d) Hardware, OS support by REX/System Administrators Head nodes same level of hardware support as tape backed dCache system File servers - support 8x5 - similar to all other file servers used by CDF CDF experiment provides primary support for the dCache software

10-October-2006CDF Analysis Disk Pool - D Benjamin24 Personel Requirements (in FTE) CDF < 0.25 FTE primary support (Alexei Varganov will provide examples in next talk including his work load) Computing Division REX system administrators Twice the effort spent on production dCache head nodes (We have 2 systems now) (Small amount of effort) dCache developers < 10 hours per month for consulation

10-October-2006CDF Analysis Disk Pool - D Benjamin25 System testing Fileserver read performance Aggregate read rate ~ 800 MB/s entire system ~ 30 minute test Network limited Dccp used to move files A Varganov

10-October-2006CDF Analysis Disk Pool - D Benjamin26 30 jobs ~ 2.5hrs 100 jobs ~9 hrs Read Transfer rate (MB/s) Network limited (both tests) Read test - Job scaling Nexsan ATA beast file server MB/s A Varganov 120 MB/s 60 MB/s 120 MB/s

10-October-2006CDF Analysis Disk Pool - D Benjamin job read performance test (dccp)

10-October-2006CDF Analysis Disk Pool - D Benjamin28 Root file read performance Note: Root files are compressed on disk - (decompression takes time) Old fileserver used in test

10-October-2006CDF Analysis Disk Pool - D Benjamin29 PNFS bandwidth testing Used both dCache prestage command and root Tfile open to make measurements. Found response linear up to ~ 1500 simultaneous clients ~ 0.1 secs per Tfile open Automatically recovered when test finished (had > 2000 simultaneous clients)

10-October-2006CDF Analysis Disk Pool - D Benjamin30 Summary In order to produce excellent physics results - CDF needs project disks space dCache provides an excellent choice for CDF CDF has extensive experience with dCache Minimizes change to users code - reduce impact on users to increase their efficiency Minimizes the operational effort