Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.

Slides:



Advertisements
Similar presentations
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Advertisements

LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
– n° 1 StoRM latest performance test results Alberto Forti Otranto, Jun
1 RAL Status and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, CERN, November 2009.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Toward new HSM solution using GPFS/TSM/StoRM integration Vladimir Sapunenko (INFN, CNAF) Luca dell’Agnello (INFN, CNAF) Daniele Gregori (INFN, CNAF) Riccardo.
ASGC 1 ASGC Site Status 3D CERN. ASGC 2 Outlines Current activity Hardware and software specifications Configuration issues and experience.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Federico Ruggieri INFN-CNAF GDB Meeting 10 February 2004 INFN TIER1 Status.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
1 INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff 28 th October 2009.
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
LFC Replication Tests LCG 3D Workshop Barbara Martelli.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
CERN IT Department CH-1211 Genève 23 Switzerland t Load Testing Dennis Waldron, CERN IT/DM/DA CASTOR Face-to-Face Meeting, Feb 19 th 2009.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
1.3 ON ENHANCING GridFTP AND GPFS PERFORMANCES A. Cavalli, C. Ciocca, L. dell’Agnello, T. Ferrari, D. Gregori, B. Martelli, A. Prosperini, P. Ricci, E.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
PADME Kick-Off Meeting – LNF, April 20-21, DAQ Data Rate - Preliminary estimate Tentative setup: all channels read with Fast ADC 1024 samples, 12.
Storage & Database Team Activity Report INFN CNAF,
STORAGE EXPERIENCES AT MWT2 (US ATLAS MIDWEST TIER2 CENTER) Aaron van Meerten University of Chicago Sarah Williams Indiana University OSG Storage Forum,
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
06-Sep-07CHEP 2007, Victoria Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.
INFN Site Report R.Gomezel October 9-13, 2006 Jefferson Lab, Newport News.
LdA, VV, PPR, GD. CHEP 2007, Victoria 06-Sep-07 Reasons for the test LHC is coming  Last chance to finalize the technical issues of our setup.
a brief summary for users
status, usage and perspectives
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
WLCG IPv6 deployment strategy
Title of the Poster Supervised By: Prof.*********
Luca dell’Agnello INFN-CNAF
SuperB – INFN-Bari Giacinto DONVITO.
The Beijing Tier 2: status and plans
LCG Service Challenge: Planning and Milestones
StoRM: a SRM solution for disk based storage systems
Status of the SRM 2.2 MoU extension
Status and plans Giuseppe Lo Re INFN-CNAF 8/05/2007.
IT-DB Physics Services Planning for LHC start-up
Andrea Chierici On behalf of INFN-T1 staff
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Service Challenge 3 CERN
Bernd Panzer-Steindel, CERN/IT
Update on Plan for KISTI-GSDC
Luca dell’Agnello INFN-CNAF
Olof Bärring LCG-LHCC Review, 22nd September 2008
STORM & GPFS on Tier-2 Milan
PES Lessons learned from large scale LSF scalability tests
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX Spring 2017
ACAT 2007 April Nikhef Amsterdam
Experience with GPFS and StoRM at the INFN Tier-1
Storage resources management and access at TIER1 CNAF
Dirk Duellmann ~~~ WLCG Management Board, 27th July 2010
The LHCb Computing Data Challenge DC06
Presentation transcript:

Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF

Reasons for the test LHC is coming  Last chance to finalize the technical issues of our setup Need to validate and perform final tests of CNAF storage infrastructure  Validation of network interconnection with the farm  Validation the SAN setup  Validation of new hardware (EMC CX3-80)  Storage system tuning and performance measurement CASTOR GPFS/StoRM  Get use of Large file system with GPFS (>100 TB) Test of other storage systems (mainly for Tier2s)  Xrootd  dcache see Giacinto’s slides

Storage CNAF Implementation of 3 Storage Classes needed for LHC Disk0Tape1 (D0T1)  CASTOR  Space managed by system  Data migrated to tapes and deleted from when staging area full Disk1tape1 (D1T1)  CASTOR  Space managed by VO (i.e. if disk is full, copy fails)  Large buffer of disk with tape back end and no gc Disk1tape0 (D1T0)  GPFS/StoRM  Space managed by VO

Our setup for the test (1) 2 EMC CX3-80  ~ 160 TB raw each 2 storage processors 4 GB of RAM for SP 500GB LCFC disks 8x4 Gbps FC ports FC Director: Brocade (x1)  128x4 Gbps FC ports 24 Disk Servers: Dell SE1950 (one used only for management)  Dual core bi-processor Xeon 1.6GHz  4 GB RAM  Dual-port Qlogic 246x HBA  1 gbps uplink to the switch  new OS used (SLC bits)  Some kernel tunings to optimize network throughput Kernel: EL.cernsmp

Our setup for the test (2) 8 wn racks (~ 1100 CPU slots)  2 Gbps uplink from the rack switches  Standard OS (SLC bits)  Standard mw (LCG 3.1)  Kernel EL.cernsmp BD core switch interconnecting rack switches and disk servers switch  2x10 Gbps links from BD to servers switch  16 Gbps aggregated from BD to wn racks switches ER16 switch not used due to interoperability issues Some minor problems with LDAP servers (used for local authz)

(Final) test-bed layout BD 1 Gbps

Test phases Network validation Local throughput measurement Storage systems performance measurement Realistic analysis SRM tests (in progress)

Network validation (netperf) flux 20 Gbps uplink validationdiskservers throughput  wns

Storage systems configuration 23 I/O servers  CASTOR, GPFS, xrootd (xfs), dcache (xfs) installed on each server For CASTOR test instance of stager used For GPFS, production cluster used (but local tests on an ad hoc cluster)  56 LUNs = 56 NSDs single file system of 110 TB  4 storage groups = 4 separated SAN zones  Data and Metadata distributed across all NSDs  Each LUN has its primary and backup server GPFS

Local throughput measurement single EMC CX3-80 with xfs type of test: dd if=/dev/zero of=/var/xxxx bs=1024k File size=12GB ~ 7.2 Gbps read 2 EMC CX3-80 with GPFS (v ) 24 server with Gbit and 2x4Gbit FC connection type of test: dd if=/dev/zero of=/gpfs/file bs=1024k File size=12GB ~ 4.4 Gbps write ~ 9.6 Gbps write ~ 12 Gbps read saturation

GPFS sequential I/O throughput 5GB files (block size=64K) ~1100 CPU slots  8 wn racks  16 Gbps network available av. ~ 9.3 Gbpsav.~ 11.4 Gbps

Dcache sequential I/O throughput 5GB files (block size=64K)  Same results with 1M block size ~1100 CPU slots  8 wn racks  16 Gbps network available ~ 6.2 Gbps~ 10.3 Gbps

Xrootd sequential I/O throughput 5GB files (block size=64K) ~1100 CPU slots  8 wn racks  16 Gbps network available ~ 6.8 Gbps ~ 8 Gbps

Castor sequential I/O throughput 5GB files (block size=64K) ~1100 CPU slots  8 wn racks  16 Gbps network available ~ 6.7 Gbps~ 9 Gbps

Sequential I/O throughput summary TestTB tbt Transf. TB TB Network Disk Net Ratio Efficiency (%) Average Network Rate (GB/s) GPFS Castor Dcache Xrootd TB tbt: foreseen TB based on the number of jobs effectively running Transferred TB : TB written on disk Efficiency: percentage of effectively transferred files Average Network Rate (GB/s): mean of 5 mns. mean values TestTB tbt Transf. TB TB Network Disk Net Ratio Efficiency (%) Average Network Rate (GB/s) GPFS Castor Dcache Xrootd write read

A real analysis case: LHCb DaVinci Test Total time [s] Avarage CPU usage (%) GPFS Dcache Xrootd CASTOR shows well known problems due to LSF plug-in limitations Xrootd performances probably could be improved tuning DaVinci

Functionality tests with StoRM (1) In collaboration with LHCb CERN people Pre-production installation  StoRM service, GridFTP service and RFIO daemon reside on the same host (storm02.cr.cnaf.infn.it)  2 GPFS disk servers with ~ 4 TB of disk  No dedicated hw on CERN side and CNAF end-point also used for SRM suite tests Data transferred from CERN CASTOR  directories from LHCb bookkeeping  a bash script takes in input the directory and does: get surl, get turl at CERN Castor prepare to put globus-url-copy  Test all normal functionalities (ls, rm, mkdir, etc...)

Functionality tests with StoRM (2) (very) preliminary measurements for sustained throughput  current deployment is very simple and not optimized Total amount of data transfered: ~1.2 TB Number of files: ~2400, from 14 directories Failures: about 4% of file transfer, due to:  get surl at CERN: 28%  get turl at CERN: 13%  get a destination turl: 0.1% (NOT an error, just the timeout set too short)  globus-url-copy: 53%  Other errors: ~ 6% srm-durable-lhcb.cern.ch  storm02 storm02  lxlpus.cern.ch srm-durable-lhcb.cr.cnaf.infn.it  storm02 srm-durable-lhcb.cern.ch  storm02

for T0D1 A true preproduction deployment for StoRM needed Planned configuration evolution:  at least 4 gridftp/GPFS servers  ~ 40 TB of GPFS file systems  3 load-balanced StoRM servers First results at the end of May

Conclusions The validation of GPFS setup has been successful  GPFS, dcache and xrootd all well performing  GPFS in particular demonstrated robustness and high performances GPFS more than suitable for usage in real world  Long experience at CNAF (> 3 years) ~ 20 GPFS file systems in production at CNAF (~ 180 net TB) mounted on all farm wns administration simple (no Oracle, no LSF, intuitive admin commands etc...)  Reliable: failure of a single file server does not affect data availability  SRM interface (StoRM) has been released CASTOR showed no surprises  Well known LFS plug-in problem Upgrade end of this month ?  Issue to investigate: more stagers needed ? (CERN model) Confident to fit LHC and non LHC experiments with present model

Credits Network setup: M. Bencivenni, D. Degirolamo, R. Veraldi, S. Zani Farm setup: M. Donatelli, A. Italiano, D. Salomoni Monitoring setup: F. Rosso, D. Vitlacil Storage hw setup: A. D’apice, PP. Ricci, V. Sapunenko Storage systems setup: G. Donvito, A. Fella, F. Furano, G. Lore, V. Sapunenko,D. Vitlacil Storage systems tests: A. Carbone, L. dell’Agnello, G. Donvito, A. Fella, F. Furano, G. Lore, V. Sapunenko, V. Vagnoni Storm development team: A. Forti, L. Magnoni, R. Zappi Storm tests: E. Lanciotti, R. Santinelli, V. Sapunenko