London Tier 2 Status Report GridPP 13, Durham, 4 th July 2005 Owen Maroney, David Colling.

Slides:



Advertisements
Similar presentations
LT 2 London Tier2 Status Olivier van der Aa LT2 Team M. Aggarwal, D. Colling, A. Fage, S. George, K. Georgiou, W. Hay, P. Kyberd, A. Martin, G. Mazza,
Advertisements

London Tier2 Status O.van der Aa. Slide 2 LT 2 21/03/2007 London Tier2 Status Current Resource Status 7 GOC Sites using sge, pbs, pbspro –UCL: Central,
Workload Management Status of current activity GridPP 13, Durham, 6 th July 2005.
NorthGrid status Alessandra Forti Gridpp15 RAL, 11 th January 2006.
Deployment metrics and planning (aka Potentially the most boring talk this week) GridPP16 Jeremy Coles 27 th June 2006.
LondonGrid Site Status and Resilience Issues Duncan Rand GridPP22.
LondonGrid Status Duncan Rand. Slide 2 GridPP 21 Swansea LondonGrid Status LondonGrid Five Universities with seven GOC sites –Brunel University –Imperial.
NorthGrid status Alessandra Forti Gridpp12 Brunel, 1 February 2005.
12th September 2002Tim Adye1 RAL Tier A Tim Adye Rutherford Appleton Laboratory BaBar Collaboration Meeting Imperial College, London 12 th September 2002.
Quarterly report ScotGrid Quarter Fraser Speirs.
Tier1A Status Andrew Sansum GRIDPP 8 23 September 2003.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
NorthGrid status Alessandra Forti Gridpp13 Durham, 4 July 2005.
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
London Tier 2 Status Report GridPP 12, Brunel, 1 st February 2005 Owen Maroney.
Southgrid Status Report Pete Gronbech: February 2005 GridPP 12 - Brunel.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Quarterly report SouthernTier-2 Quarter P.D. Gronbech.
27/04/05Sabah Salih Particle Physics Group The School of Physics and Astronomy The University of Manchester
RAL PPD Site Update and other odds and ends Chris Brew.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
INFSO-RI Enabling Grids for E-sciencE Status of LCG-2 porting Stephen Childs, Brian Coghlan and Eamonn Kenny Grid-Ireland/EGEE October.
OSG Site Provide one or more of the following capabilities: – access to local computational resources using a batch queue – interactive access to local.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
LT 2 London Tier2 Status Olivier van der Aa LT2 Team M. Aggarwal, D. Colling, A. Fage, S. George, K. Georgiou, W. Hay, P. Kyberd, A. Martin, G. Mazza,
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Quarterly report ScotGrid Quarter Fraser Speirs.
CMS Report – GridPP Collaboration Meeting V Peter Hobson, Brunel University16/9/2002 CMS Status and Plans Progress towards GridPP milestones Workload management.
Organisation Management and Policy Group (MPG): Responsible for setting and policy decisions and resolving any issues concerning fractional usage, acceptable.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
Project Management Sarah Pearce 3 September GridPP21.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
GridPP Deployment Status GridPP14 Jeremy Coles 6 th September 2005.
UK middleware deployment GridPP27 - CERN 15 th September 2011 GridPP27 - CERN 15 th September 2011 Status & plans Jeremy Coles.
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
London Tier 2 Status Report GridPP 11, Liverpool, 15 September 2004 Ben Waugh on behalf of Owen Maroney.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Southgrid Technical Meeting Pete Gronbech: May 2005 Birmingham.
University of Bristol 5th GridPP Collaboration Meeting 16/17 September, 2002Owen Maroney University of Bristol 1 Testbed Site –EDG 1.2 –LCFG GridPP Replica.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
Performance analysis extracts from GridPP OC metrics report For UKI operations meeting 15 th June 2005.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
Accounting in LCG/EGEE Can We Gauge Grid Usage via RBs? Dave Kant CCLRC, e-Science Centre.
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
The status of IHEP Beijing Site WLCG Asia-Pacific Workshop Yaodong CHENG IHEP, China 01 December 2006.
18/12/03PPD Christmas Lectures 2003 Grid in the Department A Guide for the Uninvolved PPD Computing Group Christmas Lecture 2003 Chris Brew.
London Tier-2 Quarter Owen Maroney
LCG Service Challenge: Planning and Milestones
Grid Computing for the ILC
The CCIN2P3 and its role in EGEE/LCG
Presentation transcript:

London Tier 2 Status Report GridPP 13, Durham, 4 th July 2005 Owen Maroney, David Colling

4 th July 2005GridPP 13: London Tier 2 Status Brunel 2 WN LCG-2_4_0 –R-GMA and APEL installed –RH7.3 LCFG installed Additional farm being installed –SL3 –Private networked WN –16 nodes –Expected to move into production after 2_6_0 upgrade –Hoping to bring further resources over the summer –Recruiting support post with RHUL (Job offer made)

4 th July 2005GridPP 13: London Tier 2 Status Imperial College London Appointment of Mona Aggarwal to GridPP Hardware Support Post 52 CPU Torque HEP LCG-2_5_0 –RGMA and APEL installed –OS RHEL 3 IC HEP participating in SC3 as the UK CMS site –dCache SRM installed with 2.6TB storage + 6TB on order –Another 6TB on order Numerous power outages (scheduled and unscheduled) have caused availability problems London e-Science Centre -SAMGrid installed across HEP and LeSC Certified for D0 data reprocessing 186 Job Slots –SGE farm, 64bit RHEL Globus-jobmanager installed Beta version of SGE plug-in to generic information provider Firewall issues had blocked progress but this has now been resolved. Testing will start soon. –“Community of Interest” mailing list established for sites interested in SGE integration with LCG 19 subscribers from sites in UK, Italy, Spain, Germany, France, Russia

4 th July 2005GridPP 13: London Tier 2 Status Queen Mary 320 CPU Torque farm –After difficulties with Fedora 2, have moved LCG WN to SL3 –Departure of key staff member just as LCG-2_4_0 released led to manpower problems GridPP Hardware Support post filled Guiseppe Mazza start(ed) 1 st July –RGMA and APEL installed early in June.

4 th July 2005GridPP 13: London Tier 2 Status Royal Holloway Little change: 148 CPU Torque farm –LCG 2_4_0 –OS SL3 –RGMA installed Problems with APEL default installation Gatekeeper and batch server on separate nodes Little manpower available –Shared GridPP Hardware Support post with Brunel still in recruitment process Job offer made?

4 th July 2005GridPP 13: London Tier 2 Status University College London UCL-HEP 20 CPU PBS LCG-2_4_0 –OS SL3 –RGMA installed Problems with APEL default installation Separate batch server to gatekeeper UCL-CCC 88 CPU Torque LCG- 2_4_0 –OS SL3 –RGMA and APEL installed –Main cluster is SGE farm interest in putting SGE farm into LCG and integrating nodes into single farm

4 th July 2005GridPP 13: London Tier 2 Status Current site status summary SiteService nodes Worker nodes Local network connectivity Site connectivity SRMDays SFT failed Days in scheduled maintenance BrunelRH7.3 LCG Gb100MbNo2116 ImperialRHEL3 LCG Gb dCache2628 QMULSL3 LCG2.4.0 SL3 LCG Gb100MbNo4512 RHULRHEL3 LCG Gb No2229 UCL (HEP)SL3 LCG2.4.0 SL Gb No930 UCL (CCC)SL3 LCG2.4.0 SL3 LCG Gb No129 1)Local network connectivity is that to the site SE 2)It is understood that SFT failures do not always result from site problems, but it is the best measure currently available.

4 th July 2005GridPP 13: London Tier 2 Status LCG resources SiteEstimated for LCGCurrently delivering to LCG Total job slots CPU (kSI2K) Storage (TB) Total jobs slots CPU (kSI2K) Storage (TB) Brunel IC QMUL RHUL UCL Total ) The estimated figures are those that were projected for LCG planning purposes: 2) Current total job slots are those reported by EGEE/LCG gstat page.

4 th July 2005GridPP 13: London Tier 2 Status Resources used per VO over quarter (kSI2K hours) Site CPUALICEATLASBABARCMSLHCBZEUSTotal Brunel Imperial , ,263 QMUL ,69782,854 RHUL 1,1241, ,21845,261 UCL 6, ,11521,223 Total1,1439, , ,756 Data taken from APEL

4 th July 2005GridPP 13: London Tier 2 Status Expressed as a pie chart Njobs percentage numbers of jobs 51,209 according to APEL

4 th July 2005GridPP 13: London Tier 2 Status Site Experiences LCG-2_4_0 release was first “scheduled” release date –Despite a slippage of 1 week in the release (and an overlap with EGEE conference) all LT2 sites upgraded within 3 weeks Some configuration problems for a week after –Overall experience was better than the past Farms are not fully utilised –This is true of the grid as a whole –Will extend the range of VOs supported Overall improvement in Scheduled Downtime (SD) compared to previous quarter. –QMUL had manpower problems NB: Although QMUL had highest number of (SFT failure+SD) provided most actual processing power during quarter! –IC had several scheduled power outages, plus two unscheduled power failures Caused knock-on failures for sites using BDII hosted at IC IC installed dCache SRM in preparation for SC3 –Installation configuration not simple: default configuration was not suitable for most Tier 2 sites and changing from the default was hard –Some security concerns: installations not Secure by Default Coordinator- Owen Leaving in two weeks Have made an offer