Tier-1 Andrew Sansum Deployment Board 12 July 2007.

Slides:



Advertisements
Similar presentations
RAL Tier1: 2001 to 2011 James Thorne GridPP th August 2007.
Advertisements

Tier-1 Status Andrew Sansum GRIDPP18 21 March 2007.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
IPv6 testing plans 25 Jan Short term – next 6 weeks Add sites to testbed – Glasgow (DPM storage end point) – Fix DESY – Others? Is GridFTP mesh.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Tier-1 Overview Andrew Sansum 21 November Overview of Presentations Morning Presentations –Overview (Me) Not really overview – at request of Tony.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
Tier1 Site Report HEPSysMan, RAL June 2010 Martin Bly, STFC-RAL.
Tier1 - Disk Failure stats and Networking Martin Bly Tier1 Fabric Manager.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
RAL Site Report HEPiX Fall 2013, Ann Arbor, MI 28 Oct – 1 Nov Martin Bly, STFC-RAL.
Global Science experiment Data hub Center Oct. 13, 2014 Seo-Young Noh Status Report on Tier 1 in Korea.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status Tony Cass (With thanks to Miguel Coelho dos Santos & Alex Iribarren) LCG-LHCC.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
RAL Tier 1 Site Report HEPSysMan – RAL – May 2006 Martin Bly.
Tier1 Status Report Martin Bly RAL 27,28 April 2005.
RAL Site Report Martin Bly HEPiX Fall 2009, LBL, Berkeley CA.
Martin Bly RAL Tier1/A RAL Tier1/A Report HepSysMan - July 2004 Martin Bly / Andrew Sansum.
John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
Project Management Sarah Pearce 3 September GridPP21.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team 12 th CERN-Korea.
Tier1 Site Report HEPSysMan, RAL May 2007 Martin Bly.
RAL Site Report HEPiX FAll 2014 Lincoln, Nebraska October 2014 Martin Bly, STFC-RAL.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Martin Bly RAL Tier1/A Centre Preparations for the LCG Tier1 Centre at RAL LCG CERN 23/24 March 2004.
Tier1 Andrew Sansum GRIDPP 10 June GRIDPP10 June 2004Tier1A2 Production Service for HEP (PPARC) GRIDPP ( ). –“ GridPP will enable testing.
Tier1A Status Andrew Sansum 30 January Overview Systems Staff Projects.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
RAL Site Report HEPiX - Rome 3-5 April 2006 Martin Bly.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
STATUS OF KISTI TIER1 Sang-Un Ahn On behalf of the GSDC Tier1 Team WLCG Management Board 18 November 2014.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
Eygene Ryabinkin, on behalf of KI and JINR Grid teams Russian Tier-1 status report May 9th 2014, WLCG Overview Board meeting.
RAL Site Report Martin Bly SLAC – October 2005.
Your university or experiment logo here User Board Glenn Patrick GridPP20, 11 March 2008.
Criteria for Deploying gLite WMS and CE Ian Bird CERN IT LCG MB 6 th March 2007.
The RAL Tier-1 and the 3D Deployment Andrew Sansum 3D Meeting 22 March 2006.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
8 August 2006MB Report on Status and Progress of SC4 activities 1 MB (Snapshot) Report on Status and Progress of SC4 activities A weekly report is gathered.
WLCG Operations Coordination report Maria Alandes, Andrea Sciabà IT-SDC On behalf of the WLCG Operations Coordination team GDB 9 th April 2014.
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, December 2009 Stefano Zani 10/11/2009Stefano Zani INFN CNAF (TIER1 Staff)1.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
RAL Plans for SC2 Andrew Sansum Service Challenge Meeting 24 February 2005.
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
UK Status and Plans Catalin Condurache – STFC RAL ALICE Tier-1/Tier-2 Workshop University of Torino, February 2015.
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
Tier-1 Data Storage Challenges Extreme Data Workshop Andrew Sansum 20 th April 2012.
WLCG IPv6 deployment strategy
LCG Service Challenge: Planning and Milestones
HEPiX Spring 2014 Annecy-le Vieux May Martin Bly, STFC-RAL
Andrea Chierici On behalf of INFN-T1 staff
Service Challenge 3 CERN
Update on Plan for KISTI-GSDC
Luca dell’Agnello INFN-CNAF
Olof Bärring LCG-LHCC Review, 22nd September 2008
GridPP Tier1 Review Fabric
Presentation transcript:

Tier-1 Andrew Sansum Deployment Board 12 July 2007

Agenda Monitoring Deployment tools Other stuff

Staff Changes Lex Holt (Fabric Team) left in June.

Network CERN Lightpath –10Gb line to CERN working well but recently suffered a 2 day break. SuperJanet 5 –10Gb link to site –10Gb LAN on Tir-1 –Share of 2Gb through firewall –Work underway for bypass for SRM/SE traffic

Hardware Hardware operating well – very stable EU Tenders for: –>1PB disk –>2MSI2K –1PB tape (framework purchasing) –Tape drives (just beginning to start) Underway and expected to deliver in Q4. Specification of technology required is very general and we are waiting to see solutions

RAL Site x 5530 Router A OPN Router 3 x x ADS Caches CPUs + Disks CPUs + Disks CPUs + Disks CPUs + Disks 10Gb/s to CERN N x 1Gb/s 10Gb/s 5 x x RAL Tier 2 Tier 1 Oracle systems 1Gb/s to SJ4 Tier-1 LAN

CASTOR and previous releases of CASTOR: –Implemented as a shared single instance – very unreliable with missing functionality –Unable to cope with various use cases –Essentially unusable How to make things better –Improve relationship with CERN get product improvements –1 extra contractor release now deployed: –Instance planned for ATLAS/CMS/LHCB/Others –Stable –Being load tested by CMS –Promising

dCache Still running version 1.7 –Reliability reasonable Phase out had been planned for June/July but CASTOR not sufficiently advanced –Now plan to continue running dCache at least until Christmas –Will give six months warning of closure

New Machine Room Tender underway, planned completion: August M**2 can accommodate 300 racks + 5 robots 2.3MW Power/Cooling capacity (some UPS) Office accommodation for all E-Science staff Combined Heat and Power Generation (CHP) on site Not all for GRIDPP (but you get most)!

Reliability (Recent issues) RB –Continue to see: Load related issues Database size issues (need frequent cleaning) –Now running: rb01/rb02 as general RB service rb03 dedicated to Alice and LHCB –Will add more if necessary but wish to minimise work on RB and wait for WMS Top level BDII –3 servers (March) resolved timouts for a while but recurred recently –Recent upgrade to indexing version appears to have helped CE –Experienced unidentified load problem at start of June no recurrence

SL4 SL4 test service is available with a dedicated CE and a few worker nodes Expect to run both SL3 and SL4 concurrently and gradually migrate between the two. –Migration will take place as fast as experiments want –Capacity will initially be moved at experiment’s request. – Once ATLAS/LHCB and CMS are migrated we will announce a termination date of SL3 service

Grid Only Long standing milestone that Tier-1 was to offer a “Grid Only” service by the end of August Recent discussion within UB concluded that the absence of a reliable CASTOR prevented Tier-1 offering a Grid only service PMB has subsequently said that we should nevertheless move what we can to a Grid only service. (grid only job submission for example). Position statement needs to be submitted to PMB outlining what can be achieved.