CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
CASTOR Project Status CASTOR Project Status CERNIT-PDP/DM February 2000.
12. March 2003Bernd Panzer-Steindel, CERN/IT1 LCG Fabric status
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
The CERN Computer Centres October 14 th 2005 CERN.ch.
TRIUMF Site Report for HEPiX, SLAC, October 10-14,2005 TRIUMF SITE REPORT Corrie Kost Update since Hepix Spring 2005.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
INFN – Tier1 Site Status Report Vladimir Sapunenko on behalf of Tier1 staff.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Site report: CERN Helge Meinhard (at) cern ch HEPiX spring CASPUR.
IT Infrastructure Chap 1: Definition
Planning the LCG Fabric at CERN openlab TCO Workshop November 11 th 2003 CERN.ch.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
CERN - IT Department CH-1211 Genève 23 Switzerland The Tier-0 Road to LHC Data Taking CPU ServersDisk ServersNetwork FabricTape Drives.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
CERN IT Department CH-1211 Genève 23 Switzerland Introduction to CERN Computing Services Bernd Panzer-Steindel, CERN/IT.
S.Jarp CERN openlab CERN openlab Total Cost of Ownership 11 November 2003 Sverre Jarp.
CERN.ch 1 Issues  Hardware Management –Where are my boxes? and what are they?  Hardware Failure –#boxes  MTBF + Manual Intervention = Problem!
Tier1 Andrew Sansum GRIDPP 10 June GRIDPP10 June 2004Tier1A2 Production Service for HEP (PPARC) GRIDPP ( ). –“ GridPP will enable testing.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Managing the CERN LHC Tier0/Tier1 centre Status and Plans March 27 th 2003 CERN.ch.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
Gareth Smith RAL PPD RAL PPD Site Report. Gareth Smith RAL PPD RAL Particle Physics Department Overview About 90 staff (plus ~25 visitors) Desktops mainly.
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
High Availability Technologies for Tier2 Services June 16 th 2006 Tim Bell CERN IT/FIO/TSI.
Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
Computer Centre Upgrade Status & Plans Post-C5, October 11 th 2002 CERN.ch.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
The RAL Tier-1 and the 3D Deployment Andrew Sansum 3D Meeting 22 March 2006.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
26. Juni 2003Bernd Panzer-Steindel, CERN/IT1 LHC Computing re-costing for for the CERN T0/T1 center.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Power Systems with POWER8 Technical Sales Skills V1
ALICE Computing Data Challenge VI
ICEPP, University of Tokyo
IT-DB Physics Services Planning for LHC start-up
LCG 3D Distributed Deployment of Databases
LCG Deployment in Japan
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Stuart Wild. Particle Physics Group Meeting, January 2010.
Luca dell’Agnello INFN-CNAF
UK GridPP Tier-1/A Centre at CLRC
Castor services at the Tier-0
LHC Computing re-costing for
Scalable Database Services for Physics: Oracle 10g RAC on Linux
Lee Lueking D0RACE January 17, 2002
Presentation transcript:

CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch

CERN.ch 2 See  Twiki shows work in progress: –Service Level Definition - what is required –Technical Factors - components, capacity and constraints –LCG Service Co-ordination Meeting Status –The set of activites required to deliver the building blocks on which SC4 can be built  Leads to our (evolving) hardware configurations for grid servers, operational procedures and staffing  Hope it will prove useful to other sites

CERN.ch 3 Using existing buildings  Physical location –B513 »Main Computer Room, ~1,500m 2 & 1.5kW/m 2, built for mainframes in 1970, upgraded for LHC PC clusters »Second ~1,200m 2 room created in the basement in 2003 as additional space for LHC clusters and to allow ongoing operations during the main room upgrade. Cooling limited to 500W/m 2. »Contains half of tape robotics (less heat/m2). –Tape Robot building ~50m from B513 »Constructed in 2001 to avoid loss of all CERN data due to an incident in B513. Contains half of tape robotics.

CERN.ch 4 Capacity today  2000 KSi2K batch – 1100 worker nodes –Adding 2000 KSi2K December  10 STK tape silos of 6000 slots –5 interconnected silos each in two separate buildings –Physics data split –About half of slots now occupied (after media migration) – B tape drives – 30 MB/sec –200GB capacity cartridges – 6PB total  About 2 PB raw disk storage – older servers used mirrored, newer as raid.

CERN.ch 5  Activities »Physics computing services u Interactive cluster - lxplus u Batch computing - lxbatch u Data recording, storage and management u Grid computing infrastructure »Laboratory computing infrastructure u Campus networks—general purpose and technical u Home directory, & web servers (10k+ users) u Administrative computing servers

CERN.ch 6 Physics Computing Requirements  25,000k SI2K in 2008, rising to 56,000k in 2010 –2,500-3,000 boxes (multicore, blade … ?) 200W/box.  0.1W/SI2K  6,800TB online disk in 2008, 11,800TB in 2010 –1,200-1,500 boxes, –600kW-750kW  15PB of data per year –30, GB cartridges/year –Five 6,000 slot robots/year  Sustained data recording at up to 2GB/s –Over 250 tape drives and associated servers

CERN.ch 7 Tape plans  By end 2005 we will have 40 high-duty-cycle new model tape drives and matching robotics from each of IBM (3592B) and another vendor for evaluation.  Drive data rates are expected to approach 100MB/sec  Cartridge sizes are expected to approach 500GB  Cartridge costs canonical US$120 so about 25cts/GB (compared with 60 cts/GB today).  For LHC startup operations we plan on 200 drives with these characteristics.

CERN.ch 8 Grid operations servers  Hardware matched to QoS requirements – today mostly on ad-hoc older disk servers/farm PCs –Migrate immediately critical/high services to more reliable but simple mid-range servers –Evaluate high availibility solutions to be deployed by SC4 startup looking at: »FC San multiple host/disk interconnects »HA linux (automatic failover) »Logical volume replication »Application level replication »Ready to go spare hardware for less critical services (with simple operational procedures) –Objective to reach availability levels 24 by 7.

CERN.ch 9 Mid-range server building block Dual 2.8GHz Xeon, 2GB mem, 4 hot-swap 250GB disks

CERN.ch 10 Mid-range server - back  Dual gigabit ethernet, dual power supply

CERN.ch 11 Mid-range server burnin test racks  2 different vendors

CERN.ch 12 Building 513 ground floor  Space for more…

CERN.ch 13 Current Oracle RAC cluster building blocks  Fibre-channel disks/switches infrastructure

CERN.ch 14 Oracle RAC cluster  SATA2 FC attached disks and Qlogic switches

CERN.ch 15 Oracle RAC cluster - back  Qlogic HBAs in mid-range servers

CERN.ch 16 Oracle RAC cluster  Dual FC switch connections

CERN.ch 17 Who –Contract Shift Operators: 1 person 24x7 –Technician level System Administration Team »10 team members plus 3 people for machine room operations plus engineer level manager. 24 by 7 on-call. –Engineer level teams for Physics computing »System & Hardware support: approx 10FTE »Service support: approx 10FTE »ELFms software: 3FTE plus students and collaborators. u ~30FTE-years total investment since 2001