Ben Segal CERN Information Technology Division The European Data Grid Project 2nd Workshop on Linux Clusters For Super Computing NSC, Linköping,

Slides:



Advertisements
Similar presentations
CERN STAR TAP June 2001 Status of the EU DataGrid Project Fabrizio Gagliardi CERN EU-DataGrid Project Leader June 2001
Advertisements

International Grid Communities Dr. Carl Kesselman Information Sciences Institute University of Southern California.
S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
Particle physics – the computing challenge CERN Large Hadron Collider –2007 –the worlds most powerful particle accelerator –10 petabytes (10 million billion.
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
High Performance Computing Course Notes Grid Computing.
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
Grids: Why and How (you might use them) J. Templon, NIKHEF VLV T Workshop NIKHEF 06 October 2003.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
The CrossGrid project Juha Alatalo Timo Koivusalo.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
POLITEHNICA University of Bucharest California Institute of Technology National Center for Information Technology Ciprian Mihai Dobre Corina Stratan MONARC.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Workload Management Massimo Sgaravatto INFN Padova.
Exploiting the Grid to Simulate and Design the LHCb Experiment K Harrison 1, N Brook 2, G Patrick 3, E van Herwijnen 4, on behalf of the LHCb Grid Group.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Fabric Management for CERN Experiments Past, Present, and Future Tim Smith CERN/IT.
Particle Physics and the Grid Randall Sobie Institute of Particle Physics University of Victoria Motivation Computing challenge LHC Grid Canadian requirements.
Ben Segal CERN Information Technology Division The European Data Grid Project Data Mining Workshop CSC Scientific Computing - Otaniemi.
CERN TERENA Lisbon The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000
Andrew McNab - Manchester HEP - 5 July 2001 WP6/Testbed Status Status by partner –CNRS, Czech R., INFN, NIKHEF, NorduGrid, LIP, Russia, UK Security Integration.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
DataGrid Applications Federico Carminati WP6 WorkShop December 11, 2000.
A Grid Computing Use case Datagrid Jean-Marc Pierson.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
28 March 2001F Harris LHCb Software Week1 Overview of GGF1 (Global Grid Forum) and Datagrid meeting, NIKHEF, Mar 5-9 F Harris(Oxford)
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
GridPP Presentation to AstroGrid 13 December 2001 Steve Lloyd Queen Mary University of London.
The Grid approach for the HEP computing problem Massimo Sgaravatto INFN Padova
Authors: Ronnie Julio Cole David
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
…building the next IT revolution From Web to Grid…
May http://cern.ch/hep-proj-grid-fabric1 EU DataGrid WP4 Large-Scale Cluster Computing Workshop FNAL, May Olof Bärring, CERN.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
1CHEP2000 February 2000F. Gagliardi EU HEP GRID Project Fabrizio Gagliardi
Networking: Applications and Services Antonia Ghiselli, INFN Stu Loken, LBNL Chairs.
2. WP9 – Earth Observation Applications ESA DataGrid Review Frascati, 10 June Welcome and introduction (15m) 2.WP9 – Earth Observation Applications.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
LHC Computing, CERN, & Federated Identities
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
LHC Computing, SPC-FC-CC-C; H F Hoffmann1 CERN/2379/Rev: Proposal for building the LHC computing environment at CERN (Phase 1) Goals of Phase.
EC Review – 01/03/2002 – F.Carminati – Accomplishments of the project from the end user point of view– n° 1 Accomplishments of the project from the end.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
CrossGrid Workshop, Kraków, 5 – 6 Nov-2001 Distributed Data Analysis in HEP Piotr MALECKI Institute of Nuclear Physics Kawiory 26A, Kraków, Poland.
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Grid related projects CERN openlab LCG EDG F.Fluckiger
CERN, the LHC and the Grid
LHC Data Analysis using a worldwide computing grid
LHC Computing, RRB; H F Hoffmann
Presentation transcript:

Ben Segal CERN Information Technology Division The European Data Grid Project 2nd Workshop on Linux Clusters For Super Computing NSC, Linköping, Sweden October 2001

Ben Segal CERN IT/PDP 2 Acknowledgements The choice of material presented is my own, however a lot of the material has been taken from presentations made by others, notably Leanne Guy, Les Robertson and Manuel Delfino.

Ben Segal CERN IT/PDP 3 CERN - The European Organisation for Nuclear Research The European Laboratory for Particle Physics Fundamental research in particle physics Designs, builds & operates large accelerators Financed by 20 European countries SFR 950M budget - operation + new accelerators 3,000 staff 6,000 users (researchers) from all over the world Experiments conducted by a small number of large collaborations: LEP experiment (finished) : 500 physicists, 50 universities, 20 countries, apparatus cost SFR 100M LHC experiment (future ~2005) : 2000 physicists, 150 universities, apparatus costing SFR 500M

Ben Segal CERN IT/PDP 4 CERN

Ben Segal CERN IT/PDP 5 The LEP accelerator   World’s largest particle collider, ran for 11 years.   27 km circumference, 100 m underground   Counter circulating beams of electron positron bunches   Four experiments have confirmed Standard Model predictions to high precision   Maximum collision energy of 209 GeV Many questions still remain  LHC

Ben Segal CERN IT/PDP 6 LHC in the LEP Tunnel   Counter circulating beams of protons in the same beampipe.   Centre of mass collision energy of 14 TeV.   1000 superconducting bending magnets, each 13 metres long, field 8.4 Tesla.   Super-fluid Helium cooled to K World’s largest superconducting structure

Ben Segal CERN IT/PDP 7 The LHC detectors CMS ATLAS LHCb 3.5 Petabytes/year  10 9 events/year

Ben Segal CERN IT/PDP 8 Online system Multi-level trigger Filter out background Reduce data volume Online reduction 10 7 Trigger menus Select interesting events Filter out less interesting level 1 - special hardware 40 MHz (40 TB/sec) level 2 - embedded processors level 3 - PCs 75 KHz (75 GB/sec) 5 KHz (5 GB/sec) 100 Hz (100 MB/sec) Data recording & offline analysis

Ben Segal CERN IT/PDP 9 Recording: MB/s Input: GB/s Filtering: 35K SI95 Data from detector & event builder Switch raw data summary data ~ 1 Petabyte/year TB/year High speed network Event filter farm Disk and tape servers Event filter and data recording (one experiment)

Ben Segal CERN IT/PDP 10 Data Handling and Computation for Physics Analysis ( Values for 1 experiment) GB/sec Processed data GB/sec Interactive physics analysis Batch physics analysis 350K SI95 Batch physics analysis 350K SI95 detector event summary data Raw data Event reconstruction Event reconstruction Event simulation 250K SI95 Event simulation 250K SI95 Analysis objects (extracted by physics topic) Event filter (selection & reconstruction) 35K SI95 Event filter (selection & reconstruction) 35K SI95 ~100 MB/sec ~1 PB/year Thousands of scientists 200 TB/year 500 TB/year 200 MB/sec

Ben Segal CERN IT/PDP 11 LEP to LHC Each LHC experiment requires one to two orders of magnitude greater than the TOTAL capacity installed at CERN today All LEP: < 1TB/year Rate: 4MB/sec All LHC: ~ 3PB/year Alice rate: 1GB/sec

Ben Segal CERN IT/PDP 12 How much data is involved? Level 1 Rate (Hz) High Level-1 Trigger (1 MHz) High No. Channels High Bandwidth (500 Gbit/s) High Data Archive (PetaByte) LHCB KLOE HERA-B CDF II CDF H1 ZEUS UA1 LEP NA49 ALICE Event Size (bytes) ATLAS CMS

Ben Segal CERN IT/PDP 13 Characteristics of HEP computing Eventindependence Event independence Data from each collision is processed independently Mass of independent problems with no information exchange Massivedatastorage Massive data storage Modest event size: 1-10 MB Total is very large - Petabytes for each experiment. Mostlyreadonly Mostly read only Data never changed after recording to tertiary storage But is read often ! cf.. magnetic tape as an archive medium Modestfloatingpointneeds Modest floating point needs HEP computations involve decision making rather than calculation Computational requirements in SPECint95 secs

tape servers disk servers application servers Generic model of a Fabric (computing farm) to external network local network servers

Ben Segal CERN IT/PDP 15 LHC Computing fabric at CERN *Taken from the LHC computing review

CMS:1800 physicists 150 institutes 32 countries World Wide Collaboration  distributed computing & storage capacity

Ben Segal CERN IT/PDP 17 World-wide computing Two problems: Funding will funding bodies place all their investment at CERN? Geography does a geographically distributed model better serve the needs of the world-wide distributed community? No Maybe – if it is reliable and easy to use Need to provide physicists with the best possible access to LHC data irrespective of location

Ben Segal CERN IT/PDP 18 Department    Desktop CERN – Tier 0 MONARC report: Tier 1 FNAL RAL IN2P3 622 Mbps 2.5 Gbps 622 Mbps 155 mbps Tier2 Lab a Uni. b Lab c Uni. n Regional centres - a multi tier model

Ben Segal CERN IT/PDP 19 The Basic Problem - Summary   Scalability  cost  complexity  management   Thousands of processors, thousands of disks, Petabytes of data, Terabits/second of I/O bandwidth, ….   Wide-area distribution  complexity  management  bandwidth   WANs are only and will only be ~1-10% of LANs   Distribute, replicate, cache, synchronise the data   Multiple ownership, policies, ….   Integration of this amorphous collection of Regional Centres..  .. with some attempt at optimisation   Adaptability  flexibility  simplicity   We shall only know how analysis will be done once the data arrives

Ben Segal CERN IT/PDP 20 Can Grid technology be applied to LHC computing?

Ben Segal CERN IT/PDP 21 The GRID metaphor Analogous with the electrical power grid Unlimited ubiquitous distributed computing Transparent access to multi peta byte distributed data bases Easy to plug in Hidden complexity of the infrastructure Ian Foster and Carl Kesselman, editors, “The Grid: Blueprint for a New Computing Infrastructure,” Morgan Kaufmann, 1999,

Ben Segal CERN IT/PDP 22 : : E.g., Resource-specific implementations of basic services E.g., transport protocols, name servers, differentiated services, CPU schedulers, public key infrastructure, site accounting, directory service, OS bypass Resource-independent and application-independent services authentication, authorisation, resource location, resource allocation, events, accounting, remote data access, information, policy, fault detection Distributed computing toolkit Grid Fabric (Resources) Grid Services (Middleware) Application Toolkits Data- intensive applications toolkit Collaborative applications toolkit Remote Visualisation applications toolkit Problem solving applications toolkit Remote instrumentation applications toolkit Applications Chemistry Biology Cosmology High Energy Physics Environment GRID from a services view

Ben Segal CERN IT/PDP 23 What should the Grid do for you? You submit your work … … and the Grid: Finds convenient places for it to be run Organises efficient access to your data Caching, migration, replication Deals with authentication to the different sites that you will be using Interfaces to local site resource allocation mechanisms, policies Runs your jobs Monitors progress Recovers from problems Tells you when your work is complete If there is scope for parallelism, it can also decompose your work into convenient execution units based on the available resources, data distribution

Ben Segal CERN IT/PDP 24 European Data Grid -- R&D requirements Localfabric Local fabric Management of giant computing fabrics auto-installation, configuration management, resilience, self-healing Mass storage management multi-PetaByte data storage, “real-time” data recording requirement, active tape layer – 1,000s of users, uniform mass storage interface, exchange of data and metadata between mass storage systemsWide-area Workload management no central status, local access policies Data management caching, replication, synchronisation, object database model Application monitoring Note: Build on existing components such as Globus middleware Foster (Argonne) and Kesselman (University of Southern California)

Ben Segal CERN IT/PDP 25 Managing partners UK: PPARC Italy: INFN France: CNRS Netherlands: NIKHEF ESA/ESRIN CERNIndustry IBM (UK), Compagnie des Signaux (F), Datamat (I) Associate partners Istituto Trentino di Cultura (I), Helsinki Institute of Physics / CSC Ltd (FI), Swedish Science Research Council (S), Zuse Institut Berlin (DE), University of Heidelberg (DE), CEA/DAPNIA (F), IFAE Barcelona, CNR (I), CESNET (CZ), KNMI (NL), SARA (NL), SZTAKI (HU) Other sciences KNMI(NL), Biology, Medicine Formal collaboration with USA being established European Data Grid partners

Ben Segal CERN IT/PDP 26 Preliminary programme of work Middleware WP 1 Grid Workload ManagementF. Prelz/INFN WP 2 Grid Data ManagementB. Segal/CERN WP 3 Grid Monitoring servicesR. Middleton/PPARC WP 4 Fabric ManagementO. Barring/CERN WP 5 Mass Storage ManagementJ. Gordon/PPARC Grid Fabric -- testbed WP 6 Integration TestbedF. Etienne/CNRS WP 7 Network ServicesP. Primet/CNRS Scientific applications WP 8 HEP ApplicationsF. Carminati/CERN WP 9 EO Science ApplicationsL. Fusco/ESA WP 10 Biology ApplicationsV. Breton/CNRSManagement WP 11 DisseminationM. Draoli/CNR WP 12 Project ManagementF. Gagliardi/CERN

Ben Segal CERN IT/PDP 27 Middleware : WP 1 - WP 3: wide area Workload Management WP 1 Define and implement a suitable architecture for distributed scheduling and compute resource management in a GRID environment. Maximise the global system throughput. Data management WP 2 manage and share PetaByte-scale information volumes in high-throughput production-quality grid environments. Replication/caching; Metadata mgmt.; Authentication; Query optimisation; High speed WAN data access; interface to Mass Storage Mgmt. systems. Application monitoring WP 3 Tens of thousands of components, thousands of jobs and individual users End-user - tracking of the progress of jobs and aggregates of jobs Understanding application and grid level performance

Ben Segal CERN IT/PDP 28 Middleware WP 4 - WP 5 : local fabric Fabric management WP 4 Automated installation, configuration management, system maintenance Automated monitoring and error recovery - resilience, self-healing Performance monitoring Characterisation, mapping, management of local Grid resources Mass storage management WP 5 Multi-PetaByte data storage HSM devices Uniform mass storage interface Exchange of data and metadata between mass storage systems

Ben Segal CERN IT/PDP 29 Grid fabric WP 6 - WP 7 Integration test bed WP 6 Operate prototype test beds for applications / experiments. Integrate & build successive releases of the project middleware. Demonstrate by the end of the project, test beds operating as production facilities for real end-to-end applications over large trans-European and potentially global high performance networks. Networking services WP 7 Definition and management of the network infrastructure. Monitor network traffic and performance, develop models and provide tools and data for the planning of future networks, especially concentrating on the requirements of Grids handling significant volumes of data. Deal with the distributed security aspects of Data Grid.

Ben Segal CERN IT/PDP 30 Scientific applications WP 8 - WP 10 HEP WP 8 Develop and/or adapt High Energy Physics applications (Simulation, Data Analysis, etc.) for the geographically distributed community using the functionality provided by the Data Grid, i.e. transparent access to distributed data and high performance computing facilities. Four LHC experiments involved -- requirements are similar Earth Observation WP 9 Develop Grid-aware Earth Sciences applications Facilitate access to large computational power and large distributed data files for Earth Sciences applications. Biology WP 10 High throughput for the determination of three-dimensional macromolecular structures, analysis of genome sequences. Production storage and comparison of genetic information. Retrieval and analysis of biological literature and development of a search engine for relations between biological entities.

Ben Segal CERN IT/PDP 31 Management WP 11 - WP 12 Information dissemination and exploitation WP 11 Generation of required interest necessary for the deployment of the Datagrid Project’s results Promotion of the middleware in industry projects Co-ordination of the dissemination activities undertaken by the project partners in the various European countries Industry & Research Grid Forum initiated as the main exchange place of information dissemination and potential exploitation of the Data Grid results Project management WP 12 Overall management and administration of the project Co-ordination of technical activity within the project Conflict and resource allocation resolution and external relations

Ben Segal CERN IT/PDP 32 Status Prototype work began at CERN (and in some of the collaborating institutes) before the official project start date. Globus initial installation and tests done early: several problems found and corrected. Proposal to the EU submitted on May 8 th 2000; second draft submitted in September; accepted and signed December 29 (2000). Project started officially January 1 st The first Project milestone is the Month 9 integration of early middleware and Globus on to the first testbed configurations. This is taking place as we speak.

Ben Segal CERN IT/PDP 33 EU Data Grid Main Issues Project is by EU standards very large in funding and participants Management and coordination is a major challenge Coordination between national (European) and EU Data Grid programmes Coordination with US Grid activity (GriPhyN, PPDG, Globus) Coordination of the HEP and other sciences’ objectives Very high expectations already raised, could bring disappointments

Ben Segal CERN IT/PDP 34 Conclusions The scale of the computing needs of the LHC experiments is very large compared with current experiments each LHC experiment requires one to two orders of magnitude greater capacity than the total installed at CERN today We believe that the hardware technology will be there to evolve the current architecture of “commodity clusters” into large scale computing fabrics. But there are many management problems - workload, computing fabric, data, storage in a wide area distributed environment Disappointingly, solutions for local site management on this scale are not emerging from industry The scale and cost of LHC computing imposes a geographically distributed model. The Grid metaphor describes an appropriate computing model for LHC and future HEP computing.