Martin Bly RAL Tier1/A Centre Preparations for the LCG Tier1 Centre at RAL LCG CERN 23/24 March 2004.

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

UCL HEP Computing Status HEPSYSMAN, RAL,
GridPP News NeSC opening “Media” dissemination Tier 1/A hardware Web pages Collaboration meetings Nick Brook University of Bristol.
Chris Brew RAL PPD Site Report Chris Brew SciTech/PPD.
Martin Bly RAL CSF Tier 1/A RAL Tier 1/A Status HEPiX-HEPNT NIKHEF, May 2003.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
Tier1A Status Andrew Sansum GRIDPP 8 23 September 2003.
Martin Bly RAL Tier1/A RAL Tier1/A Site Report HEPiX-HEPNT Vancouver, October 2003.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Site Report HEPHY-UIBK Austrian federated Tier 2 meeting
March 27, IndiaCMS Meeting, Delhi1 T2_IN_TIFR of all-of-us, for all-of-us, by some-of-us Tier-2 Status Report.
EU funding for DataGrid under contract IST is gratefully acknowledged GridPP Tier-1A Centre CCLRC provides the GRIDPP collaboration (funded.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
RAL Site Report HEPiX Fall 2013, Ann Arbor, MI 28 Oct – 1 Nov Martin Bly, STFC-RAL.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
April 2001HEPix/HEPNT1 RAL Site Report John Gordon CLRC, UK.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
The SLAC Cluster Chuck Boeheim Assistant Director, SLAC Computing Services.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Group Computing Strategy Introduction and BaBar Roger Barlow June 28 th 2005.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
18-20 October 2004HEPiX - Brookhaven RAL Tier1/A Site Report Martin Bly HEPiX – Brookhaven National Laboratory October 2004.
RAL Tier 1 Site Report HEPSysMan – RAL – May 2006 Martin Bly.
Tier1 Status Report Martin Bly RAL 27,28 April 2005.
March 2003 CERN 1 EDG and AliEn in Prague Dagmar Adamova INP Rez near Prague.
Martin Bly RAL Tier1/A RAL Tier1/A Report HepSysMan - July 2004 Martin Bly / Andrew Sansum.
Tier1 Report Cambridge 23rd October 2006 Martin Bly.
LCG-2 Plan in Taiwan Simon C. Lin and Eric Yen Academia Sinica Taipei, Taiwan 13 January 2004.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
Tier1 Site Report HEPSysMan, RAL May 2007 Martin Bly.
19th September 2003Tim Adye1 RAL Tier A Status Tim Adye Rutherford Appleton Laboratory BaBar UK Collaboration Meeting Royal Holloway 19 th September 2003.
HEPix April 2006 NIKHEF site report What’s new at NIKHEF’s infrastructure and Ramping up the LCG tier-1 Wim Heubers / NIKHEF (+SARA)
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
Tier1 Andrew Sansum GRIDPP 10 June GRIDPP10 June 2004Tier1A2 Production Service for HEP (PPARC) GRIDPP ( ). –“ GridPP will enable testing.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
Tier1A Status Andrew Sansum 30 January Overview Systems Staff Projects.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
RAL Site Report John Gordon HEPiX/HEPNT Catania 17th April 2002.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
ClinicalSoftwareSolutions Patient focused.Business minded. Slide 1 Opus Server Architecture Fritz Feltner Sept 7, 2007 Director, IT and Systems Integration.
RAL Site Report HEPiX - Rome 3-5 April 2006 Martin Bly.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Tier1A Status Martin Bly 28 April CPU Farm Older hardware: –108 dual processors (450, 600 and 1GHz) –156 dual processor 1400MHz PIII Recent delivery:
RAL Site Report Martin Bly SLAC – October 2005.
The RAL Tier-1 and the 3D Deployment Andrew Sansum 3D Meeting 22 March 2006.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
RAL Plans for SC2 Andrew Sansum Service Challenge Meeting 24 February 2005.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
LCG Deployment in Japan
UK GridPP Tier-1/A Centre at CLRC
QMUL Site Report by Dave Kant HEPSYSMAN Meeting /09/2019
Presentation transcript:

Martin Bly RAL Tier1/A Centre Preparations for the LCG Tier1 Centre at RAL LCG CERN 23/24 March 2004

Martin Bly RAL Tier1/A Centre Outline Overview of RAL Hardware Infrastructure Staff LCG

Martin Bly RAL Tier1/A Centre RAL Tier1/A Multi-experiment site used by LHC experements, BaBar collaboration, Dzero, H1, Minos, Zeus, QCD, theory users Committed to provide resources for BaBar (disk, CPU) – the`A’ in `Tier1/A’. CPU/disk/tape resources grown over time for some years – heterogeneous cluster Key - Flexibility

Martin Bly RAL Tier1/A Centre Systems Main farm has most of CPU hardware –Job submission: OpenPBS –Scheduling: MAUI Disk storage available to batch workers via NFS AFS on all systems Installation and Update: Kickstart, YUM Monitoring: SURE, Gangila, Nagios, Yumit

Martin Bly RAL Tier1/A Centre Present Hardware - CPUs Linux –80 dual CPU 2.66GHz/533Mhz P4 HT Xeon rack mounted systems, 2GB memory, 30GB disk, 2x1Gb/s o/b NIC –156 dual CPU 1.4GHz/133MHz P3 rack mounted systems, 1GB memory, 40GB disk, 2x100Gb/s o/b NIC –47 dual CPU 1.0GHz/133MHz P3 tower systems, 1GB memory, 40GB disk, 100Gb/s o/b NIC –40 dual CPU 600MHz/100MHz P3 tower systems, 512MB memory, 12GB disk, 100Gb/s NIC –29 dual CPU 450MHz/100MHz P2 tower systems, 256MB memory, 12GB disk, 100Mb/s NIC – to be retired soon –Various tower and rack boxes for testbed systems Solaris (Babar - being phased out) –E CPUs, 4GB memory –E CPUs, 1GB memory –4 x E420 (4CPUs each, 1GB memory)

Martin Bly RAL Tier1/A Centre Present Hardware - Disk 11 Linux rack mount servers providing ~40TB IDE disk –11 dual 2.4GHz P4 HT Xeon servers with PCIx (1GB RAM), each with: –2 Infortrend IFT-6300 arrays, each with: –12 Maxtor 200GB Diamondmax Plus 9 drives per array, most configured as 11+1 spare in RAID 5 => ~2TB/array. 26 Linux rack mount servers providing ~44TB IDE disk –26 dual 1.266GHz P3 servers (1GB RAM), each with: – 2 Accusys arrays, each with: –12 Maxtor 80GB drives -1.7TB disk per server. 3 Linux tower servers providing ~4.8TB IDE disk –3 Athlon MP single processor tower servers, each with: –1 x 3ware with 8 Maxtor DiamondMax Plus 9 as RAID5 2 Linux servers providing 300Gb SCSI RAID 5 (to be deployed). Solaris server with 4.5TB 3 x Ultra10 Solaris servers (being phased out) AFS Cell – 1.3TB, AIX +Transarc – migrate to Linux + OpenAFS server during 2004.

Martin Bly RAL Tier1/A Centre Present Hardware - Tape Storagetek 9310 Powderhorn tape silo –6000 slots –8 x STK 9940B drives (200GB 30MB/sec) –Potential 1.2PB of which 550TB allocated to HEP 4 IBM 610+ servers with two FC connections and 1Gb/s networking on PCI-X –9940 drives FC connected via 2 switches for redundancy –SCSI raid 5 disk with hot spare for 1.2Tbytes cache space

Martin Bly RAL Tier1/A Centre Switch_1Switch_2 RS6000 fsc0fsc1 fsc0 9940B fsc1fsc0fsc1fsc rmt1 rmt4rmt3rmt2 rmt5-8 AAAAAAAA STK 9310 “Powder Horn” 1Gb/s network 1.2TB

Martin Bly RAL Tier1/A Centre 2004 Hardware - May 200+ dual CPU rack mount systems: –AMD Opteron or Intel P4 HT Xeon –2 or 4GB memory –large local disk (120GB) 120+Tb configured RAID5 disk in Infortrend IDE/SCSI arrays with dual CPU servers to match –SATA disks, SCSI interface 200Gb tapes as required `Same again’ later in 2004.

Martin Bly RAL Tier1/A Centre 2004 Hardware - later Finance recently approved for next three financial years – ’04/’05/’06. Hardware spend for later in 2004 expected to be of same order +- as current round (~700GBP). Split depends on cpu/disk/tape requirements of experiments. –Priorities decided by GridPP experiment board. PPARC – long term commitment to Tier1/A centre at RAL for lifetime of LHC but no financial figures beyond next three years. On the financial roadmap of the future.

Martin Bly RAL Tier1/A Centre Purchasing / QA Strategy: –To specify what is required in some detail while leaving the vendors room to move –Use GCAT if possible to reduce the field to specialist outfits –Evaluate the offerings against detailed specs –Determine !/$ (bang/buck) –Pick the tender that provides the best overall offering – which is not necessarily the cheapest!

Martin Bly RAL Tier1/A Centre Infrastructure - Network Now: –CPUs on 100Mb/s links to 24 or 32 port switches with 1Gb/s uplinks to Tier1/A backbone switches. –Disk servers on 1Gb/s links to backbone –1Gb/s link to site backbone (1Gb/s) –2*1Gb/s link to the SuperJanet (SJ4) network via the Thames Valley Network (TVN) which itself operates at 2.5Gb/s –RAL hardware firewall (NetScreen 5400) capable of 4Gb/s bi-directionally to/from LAN/WAN –RAL is on SuperJanet development network Separate 2.5Gb/s connection MB-NG project

Martin Bly RAL Tier1/A Centre Infrastucture – Network (2) Future: –Local LAN can go to 10Gb/s as required –Partner institute in UKLight facility: UKLight will connect UK research network to USA (StarLight) and Netherlands (NetherLight) via high capacity optical links Onward connectivity to CERN Multi-channel Gb/s controlled as if end-to-end path is a single switched lightpath or wavelength. Access to UKLight as extended national development network alongside SJ4 service network Integration of UKLight with LHC requirements

Martin Bly RAL Tier1/A Centre Infrastructure - Environment RAL facilities used to host several mainframes and there are 5 large machine rooms. Share with HPC factilities and RAL computing infrastructure. 450kW cooling plant based on two compressor/heat-exchanger units with third unit on standby. 600Amps in the 415V/3 phase supply is unused.

Martin Bly RAL Tier1/A Centre Infrastructure – Environment 2 Don’t expect major capacity problems for new equipment for Tier1/A in May ( cpu units and 120+Tb disk). Some work may be needed for cooling and power distribution though no major problems expected for December 2004 disk/cpu. Tier1/A should achieve its steady state capacity of around 1000 CPUs purchased over 4 years by Dec 2005, as older equipment phased out. Ongoing survey of machine room environment. Better assessment of growth limits later this year.

Martin Bly RAL Tier1/A Centre Staff Now: 12.1 staff funded as part of GridPP project. September: increase to 16.5 (GridPP2) Additional effort will be mostly for staff to support the experiments and GRID deployment.

Martin Bly RAL Tier1/A Centre LCG activities LCG-0 March 2003 –RB, SE, CE, 5xWN, BDII, MDS LCG-1 July/Aug 2003 –RB, SE, CE, 5xWN, BDII,MDS West Region MDS Many problems with RB Decommissioned Feb 2004 though MDS and BDII still running LCG-2 Jan/Feb 2004 –RB, SE, CE, 73xWN, BDII, MDS, IC, MON West Region MDS Much more stable! Real work, mostly from ALICE