Luca dell’Agnello INFN-CNAF

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Wahid Bhimji Andy Washbrook And others including ECDF systems team Not a comprehensive update but what ever occurred to me yesterday.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Michal Kwiatek, Juraj Sucik, Rafal.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
INFN – Tier1 Site Status Report Vladimir Sapunenko on behalf of Tier1 staff.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Quarterly report SouthernTier-2 Quarter P.D. Gronbech.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
Issues in Milan Two main problems (details in the next slides): – Site excluded from analysis due to corrupted installation of some releases (mainly )
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Federico Ruggieri INFN-CNAF GDB Meeting 10 February 2004 INFN TIER1 Status.
Overview of day-to-day operations Suzanne Poulat.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
Wahid, Sam, Alastair. Now installed on production storage Edinburgh: srm.glite.ecdf.ed.ac.uk  Local and global redir work (port open) e.g. root://srm.glite.ecdf.ed.ac.uk//atlas/dq2/mc12_8TeV/NTUP_SMWZ/e1242_a159_a165_r3549_p1067/mc1.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
The LHCb Italian Tier-2 Domenico Galli, Bologna INFN CSN1 Roma,
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
IHEP Computing Site Report Shi, Jingyan Computing Center, IHEP.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Farming Andrea Chierici CNAF Review Current situation.
November 28, 2007 Dominique Boutigny – CC-IN2P3 CC-IN2P3 Update Status.
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
INFN Site Report R.Gomezel November 5-9,2007 The Genome Sequencing University St. Louis.
Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
IHEP Computing Center Site Report Shi, Jingyan Computing Center, IHEP.
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
Title of the Poster Supervised By: Prof.*********
Luca dell’Agnello INFN-CNAF
GRID OPERATIONS IN ROMANIA
LCG Service Challenge: Planning and Milestones
INFN Computing infrastructure - Workload management at the Tier-1
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Service Challenge 3 CERN
Vanderbilt Tier 2 Project
Update on Plan for KISTI-GSDC
Castor services at the Tier-0
Olof Bärring LCG-LHCC Review, 22nd September 2008
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
Bernd Panzer-Steindel CERN/IT
Universita’ di Torino and INFN – Torino
Luca dell’Agnello Daniele Cesini GDB - 13/12/2017
OffLine Physics Computing
ETHZ, Zürich September 1st , 2016
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
The LHCb Computing Data Challenge DC06
Presentation transcript:

Luca dell’Agnello INFN-CNAF CNAF readiness Luca dell’Agnello INFN-CNAF Victoria – September 1 2007

Hardware resources (1) Requested resources for 2007 (nearly) all in place CPU power installed and available 1300 KSI2k pledged for LHC available (out of a total of 3000 KSI2k) Just delivered the last part of storage In production at the end of Q3 At present 320 TB for LHC instead of 500 (out of a total of 800 TB installed) Additional tape drives and tapes ordered Tenders for 2008 resources starting soon

Hardware resources (2) During 2008 CNAF computer room will be restructured to allow distribution of electric power up to 4 MW and the installation of up to 11 APC high density islands with water cooling Preliminary: a new chiller added (end of August) First step: 2 high density islands + 1 single row (56 APC racks) will be installed in a presently not used part of the computing center All newest hw will be moved in this new location Older hw will be upgraded to increase capacity and save power All older CPU resources replaced with quad-core servers All older storage systems with newer mounting 1 TB disks During 2008 will be able to install in 56 racks : CPU: 7.5 M SpecInt2K Disco: 3 PByte Tape: 4.5 PByte After the end of the upgrade (Q1 2009) space for 70 additional racks (not all installed at once) will be available to install all the needed resources for the following years (beyond 2010)

MSS Tape library presently installed (SL 5500) contains up to 1 PB on-line A larger tape library required To contain up to 10 PB on-line It will be acquired during 2008 Tendering phase started last July CASTOR installed for T1Dx SC Improvement in stability with the July upgrade from v. 2.1.1-9 to v. 2.1.3-15 Still missing features (e.g. repack) Still problems to be solved Tape recall is an issue (10% failures in CMS August “exercise”) Work-around in place and plan to implement several gc policies (September) Plan to differentiate stagers for larger experiments and disk pools basing on type of use (more hw resources needed) Aggregate access to CASTOR

Storage system (T0D1) tests (1) Large scale validation test last Spring Involving ~200 TB of disk and 24 disk-servers ~1100 CPU slots (8 wn racks) 16 Gbps network available

Storage system (T0D1) tests (2) Seq. wr. Sequential I/O throughput 5GB files (block size=64K) Real analysis case: LHCb DaVinci 1000 analysis jobs running on production farm and accessing data via dCache, Xrootd and GPFS Seq. r. Test Total time [s] Avarage CPU usage (%) GPFS 9500 36.43 Dcache 14500 22.69 Xrootd 16000 21.17

24 by 7 operations At present 24 by 7 operations in place for urgent issues (e.g. cooling malfunction, power interruption) No physical presence on site off business hours Automatic phone calls for alarms to on shift people Alarm system effective  Automatic e-mail and SMS alarms for service problems to relevant people All resources are remotely manageable (including power cycling and console access) Effort in making all services redundant (in progress ) Best effort service (up to now)

Cooperation with experiments Since the beginning people from experiments are part of our staff both LHC and non LHC experiments Main role is to ease use of T1 to experiments Testing and verifying center operability for their experiment Cope with experiment specific services (e.g. VOBOX management) Explaining particular needs for their experiment (e.g. queue length) Cooperating in resource administration (“simple” operations on batch systems for experiment queues or on CASTOR, FTS etc…) Translate issues of the users  Model proven to work Experiments without on-site support result in more latency for problem solutions Hiring of new people is on going