November 16, 2007 Dominique Boutigny – CC-IN2P3 Grids: Tools for e-Science DoSon AC GRID School.

Slides:



Advertisements
Similar presentations
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Advertisements

Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Overview of LCG-France Tier-2s and Tier-3s Frédérique Chollet (IN2P3-LAPP) on behalf of the LCG-France project and Tiers representatives CMS visit to Tier-1.
Les Les Robertson WLCG Project Leader WLCG – Worldwide LHC Computing Grid Where we are now & the Challenges of Real Data CHEP 2007 Victoria BC 3 September.
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
1 Deployment of an LCG Infrastructure in Australia How-To Setup the LCG Grid Middleware – A beginner's perspective Marco La Rosa
SICSA student induction day, 2009Slide 1 Social Simulation Tutorial Session 6: Introduction to grids and cloud computing International Symposium on Grid.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Configuring and Maintaining EGEE Production.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
Grid Applications for High Energy Physics and Interoperability Dominique Boutigny CC-IN2P3 June 24, 2006 Centre de Calcul de l’IN2P3 et du DAPNIA.
A short introduction to the Worldwide LHC Computing Grid Maarten Litmaath (CERN)
Enabling Grids for E-sciencE ENEA and the EGEE project gLite and interoperability Andrea Santoro, Carlo Sciò Enea Frascati, 22 November.
L ABORATÓRIO DE INSTRUMENTAÇÃO EM FÍSICA EXPERIMENTAL DE PARTÍCULAS Enabling Grids for E-sciencE Grid Computing: Running your Jobs around the World.
INFSO-RI Enabling Grids for E-sciencE Project Gridification: the UNOSAT experience Patricia Méndez Lorenzo CERN (IT-PSS/ED) CERN,
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
© 2006 Open Grid Forum Enabling Pervasive Grids The OGF GIN Effort Erwin Laure GIN-CG co-chair, EGEE Technical Director
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
Jürgen Knobloch/CERN Slide 1 A Global Computer – the Grid Is Reality by Jürgen Knobloch October 31, 2007.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Overview of grid activities in France in relation to FKPPL FKPPL Workshop Thursday February 26th, 2009 Dominique Boutigny.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
INFSO-RI Enabling Grids for E-sciencE Experience of using gLite for analysis of ATLAS combined test beam data A. Zalite / PNPI.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
INFSO-RI Enabling Grids for E-sciencE An overview of EGEE operations & support procedures Jules Wolfrat SARA.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
Victoria A. White Head, Computing Division, Fermilab Fermilab Grid Computing – CDF, D0 and more..
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
DIRAC Distributed Computing Services A. Tsaregorodtsev, CPPM-IN2P3-CNRS FCPPL Meeting, 29 March 2013, Nanjing.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
2 nd EGEE/OSG Workshop Data Management in Production Grids 2 nd of series of EGEE/OSG workshops – 1 st on security at HPDC 2006 (Paris) Goal: open discussion.
CC-IN2P3: A High Performance Data Center for Research Dominique Boutigny February 2011 Toward a future cooperation with Israel.
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
Grid Computing: Running your Jobs around the World
Grid Interoperability
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
A high-performance computing facility for scientific research
LHC Data Analysis using a worldwide computing grid
Pierre Girard ATLAS Visit
CC-IN2P3 Jean-Yves Nief, CC-IN2P3 HEPiX, SLAC
GRIF : an EGEE site in Paris Region
The LHCb Computing Data Challenge DC06
Presentation transcript:

November 16, 2007 Dominique Boutigny – CC-IN2P3 Grids: Tools for e-Science DoSon AC GRID School

November 16, 2007Dominique Boutigny2 Main characteristics of a Grid A grid is an architecture and a set of software tools designed to federate distributed computing resources. Resources are in principle heterogeneous Each node of the grid is administrated locally but there should be a central coordination in order to keep the system coherent An information system (even very light) should be present in order to match the computing tasks to the computing environment The underlying network is crucial A security and authorization system should be present

November 16, 2007Dominique Boutigny3 Different kind of production Grids Computing Grid Data Grid Both Computing and Data Molecular docking Medical imagery Astronomical data LHC data processing

November 16, 2007Dominique Boutigny4 Grids are a good way to increase the computing power available for a scientific community by putting resources in common Grids federate and contribute to build scientific communities Grids are often complicated to manage – A large grid requires a strong coordination between the participating sites But

November 16, 2007Dominique Boutigny5 The LHC Computing Grid LCG

November 16, 2007Dominique Boutigny6 Concorde (15 Km) Balloon (30 Km) CD stack with 1 year LHC data! (~ 20 Km) Mt. Blanc (4.8 Km) 4 LHC experiments  15 PetaByte of data per year We have got a problem with data 100 Million SpecInt2000 This is ~ 5000 today's 8 core computers  ~15 M$ Relatively easy to setup – Each CPU core is independent of each other 15 PetaByte of data per year  Today, this is ~20 M$ if you want to put them on disk And you also need to store the Monte Carlo simulation Need to store data securely for the whole life of the experiments Complicated architecture as the data have to move worldwide Each LHC contributor should be able to have access to any data

November 16, 2007Dominique Boutigny7 A Hierarchical Grid Architecture in an International Framework CC-IN2P3 FZK PIC NDGF NIKHEF ASCC Brookhaven Fermilab TRIUMF RAL CNAF T1 (11) T0 T3 (many) T2 (52) Île de France Clermont Nantes Strasbourg Marseille Lyon CC-IN2P3 Annecy

November 16, 2007Dominique Boutigny8 LCG Vs EGEE In Europe the LHC Computing Grid is based on the multidisciplinary project EGEE  Middleware  Grid operation infrastructure PilotNew The Grid was a necessity for the LHC Computing It was a very good opportunity for other disciplines EGEE is also providing a very sophisticated operational framework Monitoring Monitoring Ticketing system Ticketing system EGEE-II: 90 partners – 32 countries – 32 M€  Crucial for the success of the project

November 16, 2007Dominique Boutigny9 LCG Vs EGEE

November 16, 2007Dominique Boutigny10

November 16, 2007Dominique Boutigny11 Interoperability 3 grid infrastructures are being used for LHC Computing 3 grid infrastructures are being used for LHC Computing –EGEE in Europe –NorduGrid in Nordic Countries –OSG in the US These 3 infrastructures are now able to interoperate These 3 infrastructures are now able to interoperate –Job submission –Operation Developments on interoperability Developments on interoperability –Short term: GIN (Grid Interoperability Now) –Longer term: SAGA / JSDL etc… They are based on different middlewares Developed within the OGF framework

November 16, 2007Dominique Boutigny12 GRID Services for the LHC Computing services Computing Element (CE) Worker nodes (WN) WNWN WN WN WN WNWN WN WN WN WNWN WN WN WN SL4 Workload Management System Storage Based on SRM  dCache  Castor  Storm  DPM File Management Transfer: FTS Cataloguing: LFC Database replication 3D - Project VOMS Virtual Organization Management Specific experiment services VO Boxes Will be used for priority management

November 16, 2007Dominique Boutigny13 The LHC Optical Private Network

November 16, 2007Dominique Boutigny14 LCG and emerging countries The grid is a complex environment which is mandatory to provide the huge computing resources necessary for the LHC The grid is a complex environment which is mandatory to provide the huge computing resources necessary for the LHC –The learning curve is steep ! Complexity … But… Complexity … But… –It provides a framework in which all the data will be available for every collaborator everywhere  This is a unique opportunity for laboratories in emerging countries to fully participate to the physics analysis

November 16, 2007Dominique Boutigny15 Lightweight Grids

November 16, 2007Dominique Boutigny16 BOINC Network Main server BOINC provide a framework for a lightweight Grid targeting CPU intensive applications running on small datasets

November 16, 2007Dominique Boutigny17 BOINC / Data analysis from the giant interferometer LIGO and GEO – Search for pulsar generated gravitational waves Fast Fourier transforms are computed on many chunks of the best data taking periods.  Search for Gravitational Wave signals on directions spread on the sky  Huge combinatorial problem Use of individual PC  Big success > participants  Contribution to scientific outreach Gravitational wave detection

November 16, 2007Dominique Boutigny18 BOINC BOINC provides a framework for a lightweight Grid which is usable to federates the usage of distributed PC BOINC provides a framework for a lightweight Grid which is usable to federates the usage of distributed PC Standalone usage is possible in many domains – BOINC is already used by several teams working in Biology. Standalone usage is possible in many domains – BOINC is already used by several teams working in Biology. Certainly a way to explore, for laboratories with limited computing resources Certainly a way to explore, for laboratories with limited computing resources

November 16, 2007Dominique Boutigny19 Java Job Submission (JJS) Developed at CC-IN2P3 by Pascal Calvat Developed at CC-IN2P3 by Pascal Calvat Java Job Submission is a very simple User Interface to submit jobs on the Grid Java Job Submission is a very simple User Interface to submit jobs on the Grid –Works on MAC, Windows and Linux –Direct submission to Computing Element –Very efficient Especially for short jobsEspecially for short jobs –Includes a learning system in order to dynamically build a list of the "best" submission sites based on their response time

November 16, 2007Dominique Boutigny20 SRB an example of a data Grid Developed at San Diego Supercomputing Center

November 16, 2007Dominique Boutigny21 SRB a Data Grid middleware (1) Many scientific applications are based on data production and analysis ATAGG CATAG GCTAT AGGCC AGATT AA

November 16, 2007Dominique Boutigny22 SRB a Data Grid middleware (2) User wants the complexity to be hidden Inspired from: Put data Get data SRB Put data DB SRB Metadata Catalog DB SRB Metadata Catalog DB SRB Metadata Catalog

November 16, 2007Dominique Boutigny23 Biomedical applications using SRB Export PC (DICOM server, SRB client) MRI Siemens MAGNETOM Sonata Maestro Class 1.5 T Acquisition Control PC DICOM   push DICOM DICOM

November 16, 2007Dominique Boutigny24 The BIRN Project Biomedical Informatics Research Network Brain imagery – Study of brain diseases

November 16, 2007Dominique Boutigny25 SRB application in HEP Projet SuperNovae Factory Data acquisition in Hawai remotely controlled from France Data are exported to CC-IN2P3 and put at physicist disposal through SRB BaBar data distribution has been using SRB since several years Hundreds of TB of data has been transferred and referenced

November 16, 2007Dominique Boutigny26 Grid5000 a research grid Grid5000 is a project to build a 5000 node grid, dedicated for research on grid technologies 9 French sites are currently hosting 3166 Grid5000 nodes Sites are connected together on a 10 Gb/s backbone A booking system allows to reserve some nodes to run experiments. It is possible to install and deploy a complete software package from the OS up to the applications on all the nodes Since recently a network connection has been established between Grid5000 and the Japanese Grid NAREGI A close collaboration between Research Grids and Production Grids is essential Research Grids will develop the future software for the production grids Production Grids will provide the framework to test new developments

November 16, 2007Dominique Boutigny27 Networks and the Digital Divide (1) ICFA Standing Committee on Interregional Connectivity R. Les Cottrell and Shahryar Khan Pinger system running on 649 sites – 128 countries – 11 world regions

November 16, 2007Dominique Boutigny28 Networks and the Digital Divide (2) Behind Europe 6 Yrs: Russia, Latin America 7 Yrs: Mid-East, SE Asia 8-9 Yrs: So. Asia 11 Yrs: Cent. Asia 12 Yrs: Africa

November 16, 2007Dominique Boutigny29 The ORIENT / TEIN2 network Internet connection difficulties are often related to the "last mile problem"  Institutes local network  Institute connection to the main country backbone  etc Are often a problem Hong Kong is also Connected to GLORIAD 45 Mb/s 622 Mb/s to be upgraded to 2x2.5 Gb/s

November 16, 2007Dominique Boutigny30 Conclusions Different kind of grid systems have been presented Different kind of grid systems have been presented –They are adapted to different kind of research –They can be very light (BOINC) or much more complicated (LCG) There are different ways to do Grid computing There are different ways to do Grid computing –Can be very simple (a single User Interface) –Can be more sophisticated (by deploying a complete Grid node) But in any case the network quality is crucial ! But in any case the network quality is crucial ! –Emerging countries should put the focus on the network development Grid is nothing by itself, only scientific applications matters !