London Tier2 Status O.van der Aa. Slide 2 LT 2 21/03/2007 London Tier2 Status Current Resource Status 7 GOC Sites using sge, pbs, pbspro –UCL: Central,

Slides:



Advertisements
Similar presentations
Open Science Grid Living on the Edge: OSG Edge Services Framework Kate Keahey Abhishek Rana.
Advertisements

Marco Verlato, INFN 23 March, 2011 ISGC2011/OGF31, Taipei,Taiwan Interoperability solutions in India 1.
LT 2 London Tier2 Status Olivier van der Aa LT2 Team M. Aggarwal, D. Colling, A. Fage, S. George, K. Georgiou, W. Hay, P. Kyberd, A. Martin, G. Mazza,
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
Review of WLCG Tier-2 Workshop Duncan Rand Royal Holloway, University of London Brunel University.
Deployment metrics and planning (aka Potentially the most boring talk this week) GridPP16 Jeremy Coles 27 th June 2006.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL June 2006.
The London Tier 2 31 st Jan 2003David Colling, London Tier 21 Institutes: Brunel, Imperial, Queen Mary, Royal Holloway, University College. Core e-Science.
LCG WLCG Operations John Gordon, CCLRC GridPP18 Glasgow 21 March 2007.
Southgrid Status Pete Gronbech: 21 st March 2007 GridPP 18 Glasgow.
LondonGrid Site Status and Resilience Issues Duncan Rand GridPP22.
Your university or experiment logo here BaBar Status Report Chris Brew GridPP16 QMUL 28/06/2006.
Your university or experiment logo here What is it? What is it for? The Grid.
Partner Logo Tier1/A and Tier2 in GridPP2 John Gordon GridPP6 31 January 2003.
LondonGrid Status Duncan Rand. Slide 2 GridPP 21 Swansea LondonGrid Status LondonGrid Five Universities with seven GOC sites –Brunel University –Imperial.
Northgrid Status Alessandra Forti Gridpp24 RHUL 15 April 2010.
NorthGrid status Alessandra Forti Gridpp12 Brunel, 1 February 2005.
UCL HEP Computing Status HEPSYSMAN, RAL,
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Torsten Antoni – LCG Operations Workshop, CERN 02-04/11/04 Global Grid User Support - GGUS -
BINP/GCF Status Report Jan 2010
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
National Grid's Contribution to LHCb IFIN-HH Serban Constantinescu, Ciubancan Mihai, Teodor Ivanoaica.
Quarterly report ScotGrid Quarter Fraser Speirs.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
London Tier 2 Status Report GridPP 13, Durham, 4 th July 2005 Owen Maroney, David Colling.
UKI-SouthGrid Overview Face-2-Face Meeting Pete Gronbech SouthGrid Technical Coordinator Oxford June 2013.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
London Tier 2 Status Report GridPP 12, Brunel, 1 st February 2005 Owen Maroney.
Quarterly report SouthernTier-2 Quarter P.D. Gronbech.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
LT 2 London Tier2 Status Olivier van der Aa LT2 Team M. Aggarwal, D. Colling, A. Fage, S. George, K. Georgiou, W. Hay, P. Kyberd, A. Martin, G. Mazza,
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
Monitoring the Grid at local, national, and Global levels Pete Gronbech GridPP Project Manager ACAT - Brunel Sept 2011.
Quarterly report ScotGrid Quarter Fraser Speirs.
Overview of day-to-day operations Suzanne Poulat.
Monitoring in EGEE EGEE/SEEGRID Summer School 2006, Budapest Judit Novak, CERN Piotr Nyczyk, CERN Valentin Vidic, CERN/RBI.
Organisation Management and Policy Group (MPG): Responsible for setting and policy decisions and resolving any issues concerning fractional usage, acceptable.
GridPP3 project status Sarah Pearce 14 April 2010 GridPP24 RHUL.
Site Report BEIJING-LCG2 Wenjing Wu (IHEP) 2010/11/21.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks WMSMonitor: a tool to monitor gLite WMS/LB.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
Maarten Litmaath (CERN), GDB meeting, CERN, 2006/02/08 VOMS deployment Extent of VOMS usage in LCG-2 –Node types gLite 3.0 Issues Conclusions.
How to get started London Tier2 O. van der Aa. 16/04/2007 Running the LT2 UK HEP Grid: GridPP, One T1, Four T2 ScotGrid Durham, Edinburgh, Glasgow NorthGrid.
London Tier 2 Status Report GridPP 11, Liverpool, 15 September 2004 Ben Waugh on behalf of Owen Maroney.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Production Manager’s Report PMB Jeremy Coles 13 rd September 2004.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
BaBar Cluster Had been unstable mainly because of failing disks Very few (
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
Accounting in LCG/EGEE Can We Gauge Grid Usage via RBs? Dave Kant CCLRC, e-Science Centre.
Criteria for Deploying gLite WMS and CE Ian Bird CERN IT LCG MB 6 th March 2007.
Analysis of job submissions through the EGEE Grid Overview The Grid as an environment for large scale job execution is now moving beyond the prototyping.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
TIFR, Mumbai, India, Feb 13-17, GridView - A Grid Monitoring and Visualization Tool Rajesh Kalmady, Digamber Sonvane, Kislay Bhatt, Phool Chand,
LCG Accounting Update John Gordon, CCLRC-RAL 10/1/2007.
A Statistical Analysis of Job Performance on LCG Grid David Colling, Olivier van der Aa, Mona Aggarwal, Gidon Moont (Imperial College, London)
On behalf of D. Colling, G. Moont, M. Aggarwal
London Tier-2 Quarter Owen Maroney
Update on Plan for KISTI-GSDC
Статус ГРИД-кластера ИЯФ СО РАН.
Presentation transcript:

London Tier2 Status O.van der Aa

Slide 2 LT 2 21/03/2007 London Tier2 Status Current Resource Status 7 GOC Sites using sge, pbs, pbspro –UCL: Central, Hep –Imperial: Hep, LeSC, ICT –Queen Mary –Royal Holloway –Brunel Total –CPU: 2.6 MSI2K –Disk: 94 TB disk (DPM and dCache)

Slide 3 LT 2 21/03/2007 London Tier2 Status MoU, where are we ? For the Disk we are at 48% of what was promised But KSI2K/TB Ratio=28 ! Sept 2007 CPU Target

Slide 4 LT 2 21/03/2007 London Tier2 Status London CPU Load Usage=(Apel CPU Time)/(Potential CPU Time) Potential CPU Time=(KSI2K Online)*(hours in a month) Monthly potential=1.7 MSI2K*hours Gives an view on how well we perform wrt cpu

Slide 5 LT 2 21/03/2007 London Tier2 Status CPU Time per VO 1) Biomed stopped in dec 2) Recovered with lhcb/cms - Supporting 21 VO helps to keep you CPU busy.

Slide 6 LT 2 21/03/2007 London Tier2 Status CPU Time: Site contributions

Slide 7 LT 2 21/03/2007 London Tier2 Status What contribution in the UK Tier2

Slide 8 LT 2 21/03/2007 London Tier2 Status New resources Online In the last quarter, both Imperial and Brunel 440 KSI2K 60 TB (dCache) 208 KSI2K 6 TB (DPM) Now has a second 1Gb connection

Slide 9 LT 2 21/03/2007 London Tier2 Status New resources to come Imperial ICT shared resources: we will get 300 KSI2K out of it. –It runs pbspro. –Well use the IC-HEP SE. –What is currently there ? We have one frontend with a VM to run RHEL3/i386 for the CE installation All RPMS installed –What need to be done ? Accounting. Adapt the GIP plugins.

Slide 10 LT 2 21/03/2007 London Tier2 Status New resources to come RHUL new cluster –Will be located at 265 kSI2k of CPU. 126 TB storage. –Remotely manager but there will be staff on site that can reboot, change disks. –The existing resources will also move there. –UL-CC is the SJ5 POP.

Slide 11 LT 2 21/03/2007 London Tier2 Status New VOs NGS enabled at Imperial LeSC but –Test suite failed globus submission without queue parameter. –Does not seem to be a problem on the sge jobmanager side. Camont and Total enabled on our RB. –RB coping difficulty with the cms production.

Slide 12 LT 2 21/03/2007 London Tier2 Status Storage is our weak point. –Tune DPM installs in all London sites. –Start with the biggest sites (QMUL) Install more pools to distribute the load. Make sure we use the latest kernels. Allocate individual pools for big VOs. –Stress the SE using CMS merge jobs or ATLAS equivalent. Cross site support –Becoming more and more important. Example: help solve getting atlas data out of ucl. –Almost all sites agreed to give access to others. But the level of access is not uniform. Need to be implemented How do we handle tickets ? What do we need to improve ?

Slide 13 LT 2 21/03/2007 London Tier2 Status Every site admins have to many source of monitoring: –SAM, Gstat, CMS Dashboard, –GridLoad, LogWatches, Dirac Monitoring. Need to aggregate different sources in one place. –Nagios is a good candidate. Possibly one instance in London Example What to improve: better monitoring #Aborted Jobs Home dir full Problem solved

Slide 14 LT 2 21/03/2007 London Tier2 Status Conclusion CPU –Monthly > 1 MSI2k*hours. –Utilization around 65%. –Will get additional 565 KSI2K. Disk –We really need more focus –Tune our DPM setups –Increase our cpu/disk ratio –Test with real cms/atlas jobs Availability –Cross site support. –Integrate the monitoring tools that exists there within nagios. 16 April: LT2 Workshop at Imperial to encourage non HEP users on to the Grid.

Slide 15 LT 2 21/03/2007 London Tier2 Status Thanks to all of the Team M. Aggarwal, D. Colling, A. Chamberlin, S. George, K. Georgiou, M. Green, W. Hay, P. Hobson, P. Kyberd, A. Martin, G. Mazza, D. Rand, G. Rybkine, G. Sciacca, K. Septhon, B. Waugh, LT 2

Slide 16 LT 2 21/03/2007 London Tier2 Status Backup slide: LCG RB backlog Matching to slow A lot of jobs waiting to be matched What is the cure ? Move to the glite wms ?