GridPP: Executive Summary Tony Doyle. Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Exec 2 Summary Grid Status: Geographical.

Slides:



Advertisements
Similar presentations
An open source approach for grids Bob Jones CERN EU DataGrid Project Deputy Project Leader EU EGEE Designated Technical Director
Advertisements

D. Britton GridPP Status - ProjectMap 8/Feb/07. D. Britton08/Feb/2007GridPP Status GridPP2 ProjectMap.
Slide 1 Steve Lloyd Grid Brokering Meeting - 4 Dec 2006 GridPP Steve Lloyd Queen Mary, University of London Grid Brokering Meeting December 2006.
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Oversight Committee Meeting.
Deployment metrics and planning (aka Potentially the most boring talk this week) GridPP16 Jeremy Coles 27 th June 2006.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL June 2006.
GridPP Meeting, 28/6/06 UB Overview à Role of the UB n Represent user community within GridPP management n Request and allocate.
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Southgrid Status Pete Gronbech: 21 st March 2007 GridPP 18 Glasgow.
Tony Doyle Executive Summary, PPARC, MRC London, 15 May 2003.
Your university or experiment logo here BaBar Status Report Chris Brew GridPP16 QMUL 28/06/2006.
Welcome to the 12 th GridPP Collaboration Meeting Introduction Steve Lloyd, Chair of the GridPP Collaboration Board Report on recent meetings of: PPARC.
Project Status David Britton,15/Dec/ Outline Programmatic Review Outcome CCRC08 LHC Schedule Changes Service Resilience CASTOR Current Status Project.
Your university or experiment logo here What is it? What is it for? The Grid.
RAL Tier1: 2001 to 2011 James Thorne GridPP th August 2007.
Partner Logo Tier1/A and Tier2 in GridPP2 John Gordon GridPP6 31 January 2003.
B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
User Board - Supporting Other Experiments Stephen Burke, RAL pp Glenn Patrick.
SouthGrid Status Pete Gronbech: 12 th March 2008 GridPP 20 Dublin.
Tony Doyle GridPP2 Specification Process Grid Steering Committee Meeting, MRC, London, 18 February 2004.
1Oxford eSc – 1 st July03 GridPP2: Application Requirement & Developments Nick Brook University of Bristol ALICE Hardware Projections Applications Programme.
Slide 1 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 GridPP and the Grid for Particle Physics Steve Lloyd Queen Mary, University of London NW Grid Seminar.
GridPP2 Status Tony Doyle. OC Actions 1.GridPP TO PROVIDE DATA ON WHAT FRACTION OF THE REGISTERED USERS WERE MAKING THE GREATEST USAGE OF THE RESOURCES.
Particle physics – the computing challenge CERN Large Hadron Collider –2007 –the worlds most powerful particle accelerator –10 petabytes (10 million billion.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Oversight Committee Meeting.
Level 1 Components of the Project. Level 0 Goal or Aim of GridPP. Level 2 Elements of the components. Level 2 Milestones for the elements.
GridPP Deployment Status, User Status and Future Outlook Tony Doyle.
Tony Doyle - University of Glasgow 3 February 2005Science Committee Meeting GridPP Status Report Tony Doyle.
Tony Doyle - University of Glasgow 30 November 2005ScotGrid Phase 2 Procurement ScotGrid Procurement … a future news item 25 June 2006: ScotGrid's 4th.
Tony Doyle GridPP2 Proposal, BT Meeting, Imperial, 23 July 2003.
HEPSYSMAN Monitoring Workshop Introduction to the Day and Overview of Ganglia Pete Gronbech.
The LHC experiments AuthZ Interoperation requirements GGF16, Athens 16 February 2006 David Kelsey CCLRC/RAL, UK
Computing for LHC Dr. Wolfgang von Rüden, CERN, Geneva ISEF students visit CERN, 28 th June - 1 st July 2009.
Service Data Challenge Meeting, Karlsruhe, Dec 2, 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Plans and outlook at GridKa Forschungszentrum.
CBPF J. Magnin LAFEX-CBPF. Outline What is the GRID ? Why GRID at CBPF ? What are our needs ? Status of GRID at CBPF.
GridPP From Prototype to Production David Britton 21/Sep/06 1.Context – Introduction to GridPP 2.Performance of the GridPP/EGEE/wLCG Grid 3.Some Successes.
GridPP 12th Collaboration Meeting Networking: Current Status Robin Tasker 31 January 2005.
15 May 2006Collaboration Board GridPP3 Planning Executive Summary Steve Lloyd.
Nick Brook University of Bristol The LHC Experiments & Lattice EB News Brief overview of the expts  ATLAS  CMS  LHCb  Lattice.
S.L.LloydGridPP CB 29 Oct 2002Slide 1 Agenda 1.Introduction – Steve Lloyd 2.Minutes of Previous Meeting (23 Oct 2001) 3.Matters Arising 4.Project Leader's.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP27 15 th Sep 2011 GridPP.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
D. Britton GridPP Status - ProjectMap 22/Feb/06. D. Britton22/Feb/2006GridPP Status GridPP2 ProjectMap.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
Project Management Sarah Pearce 3 September GridPP21.
Tony Doyle - University of Glasgow 1 July 2005Oversight Committee GridPP: Executive Summary Tony Doyle.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Jeremy Coles UK LCG Operations The Geographical Distribution of GridPP Institutes Production Manager.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Computing for Particle.
11 March 2008 GridPP20 Collaboration meeting David Britton - University of Glasgow GridPP Status GridPP20 Collaboration Meeting, Dublin David Britton,
GridPP: Executive Summary Tony Doyle. Tony Doyle - University of Glasgow Oversight Committee 8 February 2007 Outline Exec 2 Summary Grid status High level.
GridPP Building a UK Computing Grid for Particle Physics Professor Steve Lloyd, Queen Mary, University of London Chair of the GridPP Collaboration Board.
Tony Doyle - University of Glasgow 8 July 2005Collaboration Board Meeting GridPP Report Tony Doyle.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
UK Tier 1 Centre Glenn Patrick LHCb Software Week, 28 April 2006.
LCG WLCG Accounting: Update, Issues, and Plans John Gordon RAL Management Board, 19 December 2006.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
INFSO-RI Enabling Grids for E-sciencE EGEE general project update Fotis Karayannis EGEE South East Europe Project Management Board.
LCG Accounting Update John Gordon, CCLRC-RAL 10/1/2007.
 Prospective Nationale sur les Grilles de Production, Paris, L'état d'avancement des grilles.
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Collaboration Board Meeting
GridPP: Executive Summary
Presentation transcript:

GridPP: Executive Summary Tony Doyle

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Exec 2 Summary Grid Status: Geographical View: GridMap High-level View: ProjectMap Topical View: CASTOR Performance Monitoring Disaster Planning Transition Point The Icemen Cometh Outline

Tony Doyle - University of Glasgow Oversight Committee 11 October is the third full year for the Production Grid More than 10,000 kSI2k and 1 Petabyte of disk storage The UK is the largest CPU provider on the EGEE Grid Total CPU used of 25 GSI2k-hours in the last year to Sept. The GridPP2 project has met 86% of its targets with 93% of the metrics within specification (up to 07Q2) The GridPP2 project has been extended by 7 months to April 2008 –The LCG (full) Grid Service is underway –The aim is to continue to improve reliability and performance The GridPP3 proposal has been approved for 3 years through to March 2011 [total cost of £29.5m] –The aim is to provide a performant service to the experiments We anticipate a challenging period especially for the support of experiment applications running on the Grid Exec 2 Summary

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 To create a UK Particle Physics Grid and the computing technologies required for the Large Hadron Collider (LHC) at CERN To place the UK in a leadership position in the international development of an EU Grid infrastructure Context

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 View can be geographical, high-level or topical VO Views cross-location Top-level View Geographical Views Federation, Partner, Site, etc. Next level of GridMaps Large-scale Federated Grid Services Infrastructure Global GridMap Application Domain GridMap Local GridMap Alert Corrective action effect Views of the Grid

Tony Doyle - University of Glasgow Oversight Committee 11 October Geographical Status A Leadership Position

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Resource Status The latest availability figures are (approx. in case of Tier-1): Tier-1Tier-2Total CPU [kSI2k] ,088 Disk [TB] ,493 Tape [TB] >800 >800 GridPP2 capacity targets met Combined effort from all Institutions

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Aim: by 2008 (full years data taking) -CPU ~100MSI2k (100,000 CPUs) -Storage ~80PB - Involving >100 institutes worldwide -Build on complex middleware being developed in advanced Grid technology projects, both in Europe (Glite) and in the USA (VDT) 1.Prototype went live in September 2003 in 12 countries 2.Extensively tested by the LHC experiments in September February ,547 CPUs, 4398 TB storage Status in Oct 2007: 245 sites, 40,518 CPUs, 24,135 TB storage Grid Status

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Resource Accounting 100,000 3GHz CPUs CPU resources at ~required levels (just in time delivery) time LHC start-up CPU Grid-accessible disk accounting being improved Grid Operations Centre

Tony Doyle - University of Glasgow Oversight Committee 11 October High-Level Status Production Grid project nearing successful completion…

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Tape orientedDisk orientedRequest oriented Castor 3. Topical Status

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Castor Experiments migration to Castor is an important milestone weekly technical meetings set up deployment of separate instances of Castor for ATLAS, CMS and LHCb The current progress, next steps and concerns of the experiments in this area are provided in the User Board report. Tier-1 CASTOR_Experiments_Technical_Issueshttp:// CASTOR_Experiments_Technical_Issues CASTOR has recently proven to be robust under test loads and early service challenge trials CASTOR ready for deployment (disk1 storage classes) Tier-1 review planned for November

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Bridging the Experiment-Grid Gap.. Availability Status 90% max 80% typical c.f. 95% T2 target 98% T1 target 80% max 70% typical c.f. ~95% target

Tony Doyle - University of Glasgow Oversight Committee 11 October Resources Accumulated EGEE CPU Usage 102,191,758 kSI2k-hours or >100 GSI2k-hours (!) Via APEL accounting UKI: 24,788,212 kSI2k-hours

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Past years CPU Usage by experiment UK Resources

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Past years CPU Usage by Region UK Resources

Tony Doyle - University of Glasgow Oversight Committee 11 October Job Slots and Use Currently ~51% which falls short of the 70% target

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Tier-1 CPU, disk and tape resources being built up according to plan 2008 procurement well underway

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 (measured by UK Tier-1 and Tier-2 for all VOs) ~90% CPU efficiency due to i/o bottlenecks Concern that this is currently ~70% at the Tier-1 Efficiency Each experiment needs to work to improve their system/deployment practice anticipating e.g. hanging gridftp connections during batch work A big issue for the Tier-2s.. A bigger issue for the Tier-1.. target

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Intervention Policy All UK sites are given flexibility to deal with stalled jobs (in order that their CPUs are occupied more fully overall) according to the following stalled job definition: Any job consuming <10 minutes CPU over a given 6 hour period (efficiency < 0.027) is considered stalled There is a recognised intervention scheme for UK sites Stalled Jobs

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 SAM site testing Performance over past 6 months to be used for Tier- 2 hardware allocations.. The metric is to be based on SAM Test Efficiency x (CPU Delivered + Disk Available)

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 SAM site testing 30/05/0727/08/07 %age success

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Experiments categorisation: Non-scalability or general failure of the Grid data transfer / placement system. Non-scalability or general failure of the Grid workload management system. Non-scalability or general failure of the metadata / bookkeeping system. Medium-term loss of data storage resources. Medium-term loss of CPU resources. Long-term loss of data or data storage resources. Long-term loss of CPU resources. Medium- or long-term loss of wide area network. Grid security incident. Mis-estimation of resource requirements. Disaster Planning

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Disaster Modes: Importance of Communication: Work in progress.. Disaster Planning

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Transition Point From UK Particle Physics perspective the Grid is the basis for computing in the 21st Century: 1.needed to utilise computing resources efficiently and securely 2.uses gLite middleware (with evolving standards for interoperation) 3.required significant investment from PPARC (STFC) – (£100m) over 10 yrs - including support from HEFCE/SFC 4.required 3 years prototype testbed development [GridPP1] 5.provides a working production system that has been running for three years in build-up to LHC data-taking [GridPP2] 6.enables seamless discovery of computing resources: utilised to good effect across the UK – internationally significant 7.not (yet) as efficient as end-user analysts require: ongoing work to improve performance 8.ready for LHC – just in time delivery 9.future operations-led activity as part of LCG, working with EGEE/EGI (EU) and NGS (UK) [GridPP3] 10.future challenge is to exploit this infrastructure to perform (previously impossible) physics analyses from the LHC (and ILC and Fact and..)

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Proposal WritingProposal DefenceProposal Approval 31 st March 2006 – PPARC Call 31 st October 2007? – Grants implemented Transition Point Planning.. Good things take time.. ~20 months Implementation

Tony Doyle - University of Glasgow Oversight Committee 11 October 2007 Security Network Monitoring Information Services Grid Data Management Storage Interfaces Workload Management Transition Point GridPP would like to thank all the middleware developers who have contributed to the establishment of the Production Grid