Bernd Panzer-Steindel CERN/IT

Slides:



Advertisements
Similar presentations
RAL Tier1: 2001 to 2011 James Thorne GridPP th August 2007.
Advertisements

Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
6/2/2015Bernd Panzer-Steindel, CERN, IT1 Computing Fabric (CERN), Status and Plans.
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Global Science experiment Data hub Center Oct. 13, 2014 Seo-Young Noh Status Report on Tier 1 in Korea.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
LHC Computing Review Recommendations John Harvey CERN/EP March 28 th, th LHCb Software Week.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
Installing, running, and maintaining large Linux Clusters at CERN Thorsten Kleinwort CERN-IT/FIO CHEP
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
STATUS OF KISTI TIER1 Sang-Un Ahn On behalf of the GSDC Tier1 Team WLCG Management Board 18 November 2014.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
Last update: 03/03/ :37 LCG Grid Technology Area Quarterly Status & Progress Report SC2 February 6, 2004.
01. December 2004Bernd Panzer-Steindel, CERN/IT1 Tape Storage Issues Bernd Panzer-Steindel LCG Fabric Area Manager CERN/IT.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
19. November 2007Bernd Panzer-Steindel, CERN/IT1 CERN Computing Fabric Status LHCC Review, 19 th November 2007.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
26. Juni 2003Bernd Panzer-Steindel, CERN/IT1 LHC Computing re-costing for for the CERN T0/T1 center.
LCG Accounting Update John Gordon, CCLRC-RAL 10/1/2007.
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
GDB Meeting 12. January Bernd Panzer-Steindel, CERN/IT 1 Mass Storage at CERN GDB meeting, 12. January 2005.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Grid Operations in Germany T1-T2 workshop 2015 Torino, Italy Kilian Schwarz WooJin Park Christopher Jung.
13 January 2004GDB Geneva, Milos Lokajicek Institute of Physics AS CR, Prague LCG regional centre in Prague
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
15.June 2004Bernd Panzer-Steindel, CERN/IT1 CERN Mass Storage Issues.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
Dynamic Extension of the INFN Tier-1 on external resources
WLCG IPv6 deployment strategy
WLCG Workshop 2017 [Manchester] Operations Session Summary
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
Xiaomei Zhang CMS IHEP Group Meeting December
LCG Service Challenge: Planning and Milestones
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
IT-DB Physics Services Planning for LHC start-up
Andrea Chierici On behalf of INFN-T1 staff
Database Services at CERN Status Update
3D Application Tests Application test proposals
Grid Deployment Area Status Report
Database Readiness Workshop Intro & Goals
GSIAF & Anar Manafov, Victor Penso, Carsten Preuss, and Kilian Schwarz, GSI Darmstadt, ALICE Offline week, v. 0.8.
Bernd Panzer-Steindel, CERN/IT
Update on Plan for KISTI-GSDC
Proposal for obtaining installed capacity
Luca dell’Agnello INFN-CNAF
Savannah to Jira Migration
The INFN TIER1 Regional Centre
Bernd Panzer-Steindel, CERN/IT
Olof Bärring LCG-LHCC Review, 22nd September 2008
Presentation at the International Symposium on Grid Computing
LHC Computing re-costing for
Grid Canada Testbed using HEP applications
LHC Data Analysis using a worldwide computing grid
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
Status report NIKHEF Willem van Leeuwen February 11, 2002 DØRACE.
ATLAS DC2 & Continuous production
Development of LHCb Computing Model F Harris
Presentation transcript:

Bernd Panzer-Steindel CERN/IT CERN resource situation 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT Disk Space some history : Nov/Dec 2003 negotiation with Elonex about the high rate of disk server problems Jan 2004 IT re-organization of the storage responsibilities Feb 2004 Delivery of 63 TB of new disk space (Elonex), ( the ordering process for this started in September) Feb/Mar 2004 heavy testing of the new nodes, to verify stability Mar 2004 Elonex agrees to repair (replace WD disks, cages, etc.) for 55 servers  this happens in bunches of nodes (all of them in production), heavy service procedure (exchange disk servers in production, test the returned ones, etc.), needs extra ~ 10 TB intermediate cache, still ongoing April 2004 The 15 TB on the LCG prototype became unavailable, because of the computer center refurbishment (back in 2 weeks), they were used for intermediate storage and the tape repack procedures new tender released for additional ~ 25 TB ( August delivery) During March – May distribution of >63 TB disk space to LHC experiments and running fixed target experiments 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT Disk space allocations The disk resource published in the GDB allocation round in September 2003 was ~100 TB for the LHC experiments at CERN 60 TB new purchase + 25 TB prototype contribution + 15 TB existing space from the experiments (installed end 2003 : ALICE (5.6TB), ATLAS (12.6TB), CMS (10TB), LHCb (5.5TB) that was too optimistic, due to budget constraints and all the problems in Q1+Q2 2004. I am now assuming a hopefully pessimistic 50 TB for 2004. Building the DC disk buffers, current status March-April : CMS : 23 TB ALICE : 25 TB May : LHCb : 6 TB ATLAS : 6 TB  10 TB (end May) CMS : 8 TB (reduced from 23) June – July : ATLAS : 15 – 20 TB LHCb : 6 TB CMS : 8 TB ALICE : 25 TB 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT Issues Disk space ‘re-distribution after the end of this years DCs (summer) ?? ALICE 8 TB, ATLAS 8 TB, CMS 8 TB, LHCb 6TB + 20 TB tests ???? e.g. large amount of disk space for the new stager test, the CDR/export buffer for 2008 needs to be tested on a large scale ????  needs some discussion The DC mode will change now into a more continuous production and test mode  dynamic exchange of extra space between experiments will become more difficult by the end of May the disk server replacement procedure will be finished and the also the protoype/openlab disk space will be available in a reliable manner.  upcoming computing DCs 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT CPU server distribution HE = High-End, LE = Low-End, HE = 2 * LE performance 650 HE + 480 LE nodes Lxbatch 220 LE nodes LCG testbeds 120 HE nodes prototype (WAN tests, ALICE-IT DC, new technology) (shared in Lxbatch from time to time) 40 LE nodes EGEE testbeds 30 HE nodes GRID services (RLS, SE, UI, GridFTP, etc.) 20 LE nodes DB and ARDA -------------------- 1560 nodes = 1160 production + 400 testbeds 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT CPU resources (I) renumbering process in parallel to production in units of racks ( = 44 nodes = 70 KSI2000) (outgoing network requirement)  drain the LSF queues, re-installation with GRID software packages, renumbering, change of databases and router configurations 200 nodes in February, now we have 500 nodes (the most powerful nodes first) LCG usage not very high (configuration issues, experiment usage)  system is shared between ‘standard’ Lxbatch and LCG-2  25 nodes always dedicated to LCG-2, plus the rest (total 200) with higher priority for LCG-2 jobs , mostly ‘short’ jobs from ‘standard’ Lxbatch 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT CPU resources (II) March-April : 250 nodes ( 400 KSI2000)dedicated to CMS (average efficiency ~ 30 %  dedication was very good for the experiment, fast restart after problem resolution, better suited for DC characteristics but bad for overall IT resource utilization shared  dedication adjustment/compromise Alice using shared queues (LCG-2  Lxbatch) ~ 100 nodes equivalent 340 nodes fro ATLAS online from low end Lxbatch and prototype new CPU server tender 200-400 nodes ( August delivery) May-July : LHCb (390 KSI2000), ALICE (180 KSI2000), ATLAS (170 KSI2000) fixed target ( ~230KSI2000), baseline LHC + others (200 KSI2000) available = ~ 1300 KSI2000 (efficiency in shared mode = ~ 90 %) CERN resources published in the last GDB allocation round = 700 KSI2000 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT CPU Resource Usage New purchase Folded with prototype capacity July LHC experiment plans fro Q3/Q4 still under discussion 04. May 2004 Bernd Panzer-Steindel CERN/IT

Bernd Panzer-Steindel CERN/IT Issues Complicated resource shuffling (changing demands, disk server problems, center refurbishment, budget constraints, delivery schedules, etc.) we are always working under 100% usage conditions, very little room to maneuver DCs are becoming productions, also moving into the continuous mode balance of ‘physics’ needs versus ‘computing’ needs (tests, architecture, software, etc.) Balance between production farm and testbeds (resource distribution) Balance between resource dedication and sharing 04. May 2004 Bernd Panzer-Steindel CERN/IT