Download presentation
Presentation is loading. Please wait.
Published byIsabella Wells Modified over 10 years ago
1
Slide 1 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 GridPP and the Grid for Particle Physics Steve Lloyd Queen Mary, University of London NW Grid Seminar May 2006
2
Slide 2 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 19 UK Universities, CERN and CCLRC (RAL & Daresbury) Funded by PPARC: GridPP1 2001-2004 From Web to Grid GridPP2 2004-2007 From Prototype to Production GridPP3 2007-2011 (proposed) From Production to Exploitation Who are GridPP? Developed a working, highly functional Grid
3
Slide 3 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Why? - The CERN LHC 4 Large Experiments The worlds most powerful particle accelerator – Starting 2007
4
Slide 4 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 ALICE - heavy ion collisions, to create quark-gluon plasmas - 50,000 particles in each collision LHCb - to study the differences between matter and antimatter - detect over 100 million b and b-bar mesons each year ATLAS - General purpose - Origin of mass - Supersymmetry - 2,000 scientists from 34 countries CMS - General purpose - 1,800 scientists from over 150 institutes The Experiments
5
Slide 5 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 7,000 tonnes 42m long 22m wide 22m high 2,000 Physicists 150 Institutes 34 Countries The ATLAS Detector (About the height of a 5 storey building)
6
Slide 6 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006
7
Slide 7 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Starting from this event… We are looking for this signature Selectivity: 1 in 10 13 Like looking for 1 person in a thousand world populations Or for a needle in 20 million haystacks! The LHC Data Challenge ~100,000,000 electronic channels 800,000,000 proton- proton interactions per second 0.0002 Higgs per second 10 PBytes of data a year (10 Million GBytes = 14 Million CDs)
8
Slide 8 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Solution – Build a Grid Share more than information Efficient use of resources at many institutes Leverage over other sources of funding Data, computing power, applications Join local communities Challenges: share data between thousands of scientists with multiple interests link major and minor computer centres ensure all data accessible anywhere, anytime grow rapidly, yet remain reliable for more than a decade cope with different management policies of different centres ensure data security be up and running routinely by 2007
9
Slide 9 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Middleware is Everything MIDDLEWARE CPU Disks, CPU etc PROGRAMS OPERATING SYSTEM Word/Excel Email/Web Your Program Games CPU Cluster User Interface Machine CPU Cluster CPU Cluster Resource Broker Information Service Single PC Grid Disk Server Your Program Middleware is the Operating System of a distributed computing system Replica Catalogue Bookkeeping Service
10
Slide 10 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 GridPP Middleware Development Workload Management Storage Interfaces Network Monitoring SecurityInformation Services Grid Data Management
11
Slide 11 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 International Context LHC Computing Grid (LCG) Grid Deployment Project for LHC EU Enabling Grids for e-Science (EGEE) 2004-2008 Grid Deployment Project for all disciplines GridPP LCG EGEE GridPP is part of EGEE and LCG (currently the largest Grid in the world) UK National Grid Service UKs core production computational and data Grid Open Science Grid (USA) Science applications from HEP to biochemistry Nordugrid (Scandinavia) Grid Research and Development collaboration
12
Slide 12 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 What you need to use the Grid 1. Get a digital certificate (UK Certificate Authority) 2. Join a Virtual Organisation (VO) 3. Get access to a local User Interface Machine (UI) and copy your files and certificate there Authentication – who you are Authorisation – what you are allowed to do 4. Write some Job Description Language (JDL) and scripts to wrap your programs ############# HelloWorld.jdl ################# Executable = "/bin/echo"; Arguments = "Hello welcome to the Grid "; StdOutput = "hello.out"; StdError = "hello.err"; OutputSandbox = {"hello.out","hello.err"}; #########################################
13
Slide 13 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 How it works Job Description UI Machine Resource Broker Input Sandbox Script you want to run Other files (Job Options, Source...) Storage Element Compute Element Storage Element Grid Proxy Certificate Output Sandbox Output files (Plots, Logs...) Input Data Output Data Job Output Sandbox
14
Slide 14 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Tier Structure Tier 0 Tier 1 National centres Tier 2 Regional groups Institutes Workstations Offline farm Online system CERN computer centre RAL,UK ScotGridNorthGridSouthGridLondon FranceItalyGermanyUSA GlasgowEdinburghDurham Useful model for Particle Physics but not necessary for others
15
Slide 15 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 UK Tier-1 Centre at RAL High quality data services National and International Role UK focus for International Grid development 1000 Dual CPU 200 TB Disk 220 TB Tape (Capacity 1PB) Grid Operations Centre
16
Slide 16 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 UK Tier-2 Centres ScotGrid Durham, Edinburgh, Glasgow NorthGrid Daresbury, Lancaster, Liverpool, Manchester, Sheffield SouthGrid Birmingham, Bristol, Cambridge, Oxford, RAL PPD London Brunel, Imperial, QMUL, RHUL, UCL Mostly funded by HEFCE
17
Slide 17 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 The LCG Grid Status Worldwide 182 Sites 23,438 CPUs 9.2 PB Disk 2,200 Years of CPU time UK 21 Sites 4,482 CPUs 180 TB Disk 593 Years of CPU time
18
Slide 18 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Who else can use a Grid? Astronomy Healthcare Bioinformatics Gaming Engineering Commerce
19
Slide 19 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 UK contributes to EGEE's battle with malaria BioMed Successes/Day 1107 Success % 77% WISDOM (Wide In Silico Docking On Malaria) The first biomedical data challenge for drug discovery, which ran on the EGEE grid production service from 11 July 2005 until 19 August 2005. GridPP resources in the UK contributed ~100,000 kSI2k-hours from 9 sites Number of Biomedical jobs processed by country Normalised CPU hours contributed to the biomedical VO for UK sites, July-August 2005
20
Slide 20 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 "GridPP has been developed to help answer questions about the conditions in the Universe just after the Big Bang," said Professor Keith Mason, head of the Particle Physics and Astronomy Research Council (PPARC). "But the same resources and techniques can be exploited by other sciences with a more direct benefit to society."
21
Slide 21 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 GridPP and Industry Use of our Grid Security tools GridSite Secure Web Toolkit R-GMA Information System APEL Accounting tools Help getting PPARC funding What We Have to Offer: Current Involvement: HP are sponsoring a joint project with GridPP at Bristol. GridPP has a collaboration with IBM through ScotGrid and R-GMA. Specific sites also have close relationships with various industrial suppliers.
22
Slide 22 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Security Tools and Gridsite Set of extensions to Apache Webserver Gridsite identifies users to websites by their digital certificates GridSite brings Grid Security to Websites Security is Paramount UK Certificate Authority Users must have digital User Certificates (can be revoked) Users must be part of a Virtual Organisation (VOMS Management) All Grid computers must have Host Certificates (can be revoked) Security policies agreed between sites Allows remote editing by authorised users without login access to host servers GridSiteWiki is an extension GridSite is open source (http://www.gridsite.org/)
23
Slide 23 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Future Challenges (Ex-)Concorde (15 km) CD stack with 1 year LHC data (~ 20 km) We are here (4 km) Scaling to full size ~20,000 100,000 CPUs Stability, Robustness etc Security (Hackers Paradise!) Sharing resources (in RAE environment!) International Collaboration Increased Industrial take-up Spread beyond Science Continued funding beyond start of LHC!
24
Slide 24 of 24 Steve Lloyd NW Grid Seminar - 11 May 2006 Further Info … http://www.gridpp.ac.uk RSS News feed
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.