LHC: An example of a Global Scientific Community Sergio Bertolucci CERN 5 th EGEE User Forum Uppsala, 14 th April 2010.

Slides:



Advertisements
Similar presentations
CERN STAR TAP June 2001 Status of the EU DataGrid Project Fabrizio Gagliardi CERN EU-DataGrid Project Leader June 2001
Advertisements

Computing for LHC Dr. Wolfgang von Rüden, CERN, Geneva ISEF students visit CERN, 28 th June - 1 st July 2009.
Accelerating Science and Innovation Welcome to CERN.
Welcome to CERN Accelerating Science and Innovation 2 nd March 2015 – Bidders Conference – DO-29161/EN.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Welcome to CERN Research Technology Training Collaborating.
Project Status Report Ian Bird Computing Resource Review Board 30 th October 2012 CERN-RRB
Accelerating Science and Innovation. Science for Peace CERN was founded in 1954 as a Science for Peace Initiative by 12 European States Member States:
13 October 2014 Eric Grancher, head of database services, CERN IT Manuel Martin Marquez, data scientist, CERN openlab.
Massive Computing at CERN and lessons learnt
Resources and Financial Plan Sue Foffano WLCG Resource Manager C-RRB Meeting, 12 th October 2010.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Emmanuel Tsesmelis 2 nd CERN School Thailand 2012 Suranaree University of Technology.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
Experience with the WLCG Computing Grid 10 June 2010 Ian Fisk.
Welcome – Benvenuti Carlo Verdone to Accelerating Science and Innovation to Accelerating Science and Innovation.
Introduction to CERN Dr. Wolfgang von Rüden, CERN, Geneva Thanks to F. Briard for providing some of the slides June 2010.
Petabyte-scale computing for LHC Ian Bird, CERN WLCG Project Leader WLCG Project Leader ISEF Students 18 th June 2012 Accelerating Science and Innovation.
Rackspace Analyst Event Tim Bell
Notur: - Grant f.o.m is 16.5 Mkr (was 21.7 Mkr) - No guarantees that funding will increase in Same level of operations maintained.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
Data Logistics in Particle Physics Ready or Not, Here it Comes… Prof. Paul Sheldon Vanderbilt University Prof. Paul Sheldon Vanderbilt University.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
Ian Bird LCG Project Leader OB Summary GDB 10 th June 2009.
Ian Bird LCG Project Leader Project status report WLCG LHCC Referees’ meeting 16 th February 2010.
Progress in Computing Ian Bird ICHEP th July 2010, Paris
Jürgen Knobloch/CERN Slide 1 A Global Computer – the Grid Is Reality by Jürgen Knobloch October 31, 2007.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
…building the next IT revolution From Web to Grid…
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
European Organization for Nuclear Research. Topics Founding Organization The accelerator Experiments Practical applications Video?
WelcomeWelcome CSEM – CERN Day 23 rd May 2013 CSEM – CERN Day 23 rd May 2013 to Accelerating Science and Innovation to Accelerating Science and Innovation.
CERN as a World Laboratory: From a European Organization to a global facility CERN openlab Board of Sponsors July 2, 2010 Rüdiger Voss CERN Physics Department.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Project Status Report Ian Bird Computing Resource Review Board 20 th April, 2010 CERN-RRB
Introduction to CERN and Grid Computing Dr. Wolfgang von Rüden CERN, Geneva HP ProCurve event CERN, 20 February 2008.
LHC Computing, CERN, & Federated Identities
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Frédéric Hemmer IT Department 26 th January 2010 Visit of Michael Dell 1 Frédéric.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 1 st March 2011 Visit of Dr Manuel Eduardo Baldeón.
The Mission of CERN  Push back  Push back the frontiers of knowledge E.g. the secrets of the Big Bang …what was the matter like within the first moments.
Germany and CERN / June 2009Germany and CERN | May Welcome - Willkommen CERN: to CERN: Accelerating Science and Innovation Professor Wolfgang A.
WLCG after 1 year with data: Prospects for the future Ian Bird; WLCG Project Leader openlab BoS meeting CERN4 th May 2011.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
WLCG: The 1 st year with data & looking to the future WLCG: Ian Bird, CERN WLCG Project Leader WLCG Project LeaderLCG-France; Strasbourg; 30 th May 2011.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Collaborative Research Projects in Australia: High Energy Physicists Dr. Greg Wickham (AARNet) Dr. Glenn Moloney (University of Melbourne) Global Collaborations.
WLCG – Status and Plans Ian Bird WLCG Project Leader openlab Board of Sponsors CERN, 23 rd April 2010.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
5-minutes tour of CERN (based on official CERN slides) 5-minutes tour of CERN (based on official CERN slides) Christian Joram / CERN EIROfrum Topical Workshop.
The CMS Experiment at LHC
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
The 5 minutes tour of CERN The 5 minutes race of CERN
Grid site as a tool for data processing and data analysis
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
LHC Computing Grid Status of Resources Financial Plan and Sue Foffano
Long-term Grid Sustainability
The 5 minutes tour of CERN The 5 minutes race of CERN
Dagmar Adamova, NPI AS CR Prague/Rez
The LHC Computing Grid Visit of Her Royal Highness
Input on Sustainability
CERN, the LHC and the Grid
LHC Data Analysis using a worldwide computing grid
Overview & Status Al-Ain, UAE November 2007.
The LHC Computing Grid Visit of Professor Andreas Demetriou
CERN: from fundamental sciences to daily applications
Presentation transcript:

LHC: An example of a Global Scientific Community Sergio Bertolucci CERN 5 th EGEE User Forum Uppsala, 14 th April 2010

14th April 2010Sergio Bertolucci, CERN 2 LEP LHC Our universe is expanding and cooling down and has been doing so for approximately 13.7 billion years

> 95% UNKNOWN STUFF OUT THERE > 95% UNKNOWN STUFF OUT THERE Black hole Sergio Bertolucci, CERN3 14th April 2010

Why do particles have mass? – Newton could not explain it - and neither can we… What is 96% of the Universe made of? – We only observe 4% of it! Why is there no antimatter left in the Universe? – Nature should be symmetrical What was matter like during the first second of the Universe, right after the "Big Bang"? – A journey towards the beginning of the Universe gives us deeper insight Sergio Bertolucci, CERN4 Fundamental Physics Questions 14th April 2010 The Large Hadron Collider (LHC), allows us to look at microscopic big bangs to understand the fundamental laws of nature

fundamental research and discoveries technological innovation training and education bringing the world together Sergio Bertolucci, CERN6 CERN stands for over 50 years of… 14th April Rebuilding Europe First meeting of the CERN Council 1980 East meets West Visit of a delegation from Beijing 2010 Global Collaboration The Large Hadron Collider involves over 100 countries

The world’s most powerful accelerator: LHC – A 27 km long tunnel filled with high-tech instruments – Equipped with thousands of superconducting magnets – Accelerates particles to energies never before obtained – Produces particle collisions creating microscopic “big bangs” Very large sophisticated detectors – Four experiments each the size of a cathedral – Hundred million measurement channels each – Data acquisition systems treating Petabytes per second Significant computing to distribute and analyse the data – A Computing Grid linking ~200 computer centres around the globe – Sufficient computing power and storage to handle 15 Petabytes per year, making them available to thousands of physicists for analysis Global collaborations essential at all stages Sergio Bertolucci, CERN6 CERN’s Tools 14th April 2010

Sergio Bertolucci, CERN7 14th April 2010

Sergio Bertolucci, CERN8 LHC is in operation! 14th April 2010

From this (October 2008)... Collateral damage: magnet displacements Collateral damage: ground supports 9Sergio Bertolucci, CERN

... To this (Nov 2009) 10Sergio Bertolucci, CERN

11... And now at 7 TeV

ATLAS CMS Overall weight (tons) Diameter 22 m 15 m Length 46 m 22 m Solenoid field 2 T 4 T ATLAS superimposed to the 5 floors of building 40 CMS ATLAS Scale of ATLAS and CMS? Sergio Bertolucci, CERN12 14th April 2010

Sergio Bertolucci, CERN13 14th April 2010

Sergio Bertolucci, CERN14 14th April 2010 And similar numbers for the other 3 experiments

Sergio Bertolucci, CERN 15 Collisions at the LHC: summary 14th April 2010

pp collisions at 14 TeV at cm -2 s -1 How to extract this: (Higgs  4 muons) From this: With: 20 proton-proton collisions overlap And this repeats every 25 ns… A very difficult environment … Z  at LEP (e+e-) 16Sergio Bertolucci, CERN

17 The LHC Computing Challenge 14th April 2010  Signal/Noise: (10 -9 offline)  Data volume High rate * large number of channels * 4 experiments  15 PetaBytes of new data each year  Compute power Event complexity * Nb. events * thousands users  200 k of (today's) fastest CPUs  45 PB of disk storage  Worldwide analysis & funding Computing funding locally in major regions & countries Efficient analysis everywhere  GRID technology

A distributed computing infrastructure to provide the production and analysis environments for the LHC experiments Managed and operated by a worldwide collaboration between the experiments and the participating computer centres The resources are distributed – for funding and sociological reasons Our task is to make use of the resources available to us – no matter where they are located – We know it would be simpler to put all the resources in 1 or 2 large centres – This is not an option... today Sergio Bertolucci, CERN18 WLCG – what and why? 14th April 2010

LCG was set up as a project in 2 phases: – Phase I – Development & planning; prototypes End of this phase the computing Technical Design Reports were delivered (1 for LCG and 1 per experiment) – Phase II – – Deployment & commissioning of the initial services Program of data and service challenges During Phase II, the WLCG Collaboration was set up as the mechanism for the longer term: – Via an MoU – signatories are CERN and the funding agencies – Sets out conditions and requirements for Tier 0, Tier 1, Tier 2 services, reliabilities etc (“SLA”) – Specifies resource contributions – 3 year outlook Sergio Bertolucci, CERN19 (w)LCG – Project and Collaboration 14th April 2010

De-FZK US-FNAL Ca- TRIUMF NDGF CERN Barcelona/PIC Lyon/CCIN2P3 US-BNL UK-RAL Taipei/ASGC 14th April 2010Sergio Bertolucci, CERN 20 Today we have 49 MoU signatories, representing 34 countries: Australia, Austria, Belgium, Brazil, Canada, China, Czech Rep, Denmark, Estonia, Finland, France, Germany, Hungary, Italy, India, Israel, Japan, Rep. Korea, Netherlands, Norway, Pakistan, Poland, Portugal, Romania, Russia, Slovenia, Spain, Sweden, Switzerland, Taipei, Turkey, UK, Ukraine, USA. WLCG Today Tier 0; 11 Tier 1s; 61 Tier 2 federations (121 Tier 2 sites) Amsterdam/NIKHEF-SARA Bologna/CNAF

Running increasingly high workloads: – Jobs in excess of 650k / day; Anticipate millions / day soon – CPU equiv. ~100k cores Workloads are: – Real data processing – Simulations – Analysis – more and more (new) users Data transfers at unprecedented rates  next slide Sergio Bertolucci, CERN21 Today WLCG is: e.g. CMS: no. users doing analysis

Sergio Bertolucci, CERN22 Data transfers Final readiness test (STEP’09) Preparation for LHC startupLHC physics data Nearly 1 petabyte/week 2009: STEP09 + preparation for data Castor traffic last week: > 4 GB/s input > 13 GB/s served Castor traffic last week: > 4 GB/s input > 13 GB/s served Real data – from 30/3

Sergio Bertolucci, CERN23 Fibre cut during STEP’09: Redundancy meant no interruption Fibre cut during STEP’09: Redundancy meant no interruption

MoU defines key performance and support metrics for Tier 1 and Tier 2 sites – Reliabilities are an approximation for some of these – Also metrics on response times, resources, etc. The MoU has been an important tool in bringing services to an acceptable level Sergio Bertolucci, CERN24 Service quality: defined in MoU

Focus on real and continuous production use of the service over several years (simulations since 2003, cosmics) Data and Service challenges to exercise all aspects of the service – not just for data transfers, but workloads, support structures etc. Challenges – SC1  December 2004 – SC2  March 2005 – SC3  July2005 Testing with special emphasis on Data Management Goals largely exceeded for the T2 sites, service reliability and sustained transfer rates – SC4  June 2006 Offline data processing requirements can be handled by the Grid to the nominal LHC data rate Large participation of T2 sites, all T1 sites were in Required transfer rates (disk-tape) achieved and in some cases exceeded – CCRC’08  March + June 2008 Measurement of the readiness of the Grid services and operations before real data takin All experiments simultaneously stressing the WLCG infrastructure in close to real conditions Experiments running their Full Dress Rehearsals and scheduling key periods together with the CCRC’08 challenge – STEP’09  May 2009 Stress and scale testing of all experiment workloads including tape recall and massive end user analysis Sergio Bertolucci, CERN25 Success with real data because:

Has meant very rapid data distribution and analysis – Data is processed and available at Tier 2s within hours! Sergio Bertolucci, CERN26 Readiness of the computing CMS ATLAS LHCb

Sergio Bertolucci, CERN27 And physics output...

CERN and the HEP community have been involved with grids from the beginning Recognised as a key technology for implementing the LHC computing model HEP work with EC-funded EDG/EGEE in Europe, iVDGL/Grid3/OSG etc. in US has been of clear mutual benefit – Infrastructure development driven by HEP needs – Robustness needed by WLCG is benefitting other communities – Transfer of technology from HEP Ganga, AMGA, etc used by many communities now Sergio Bertolucci, CERN28 Grids & HEP: Common history 14th April 2010

LHC, the experiments, & computing have taken ~20 years to build and commission They will run for at least 20 years We must be able to rely on long term infrastructures – Global networking – Strong and stable NGIs (or their evolution) That should be eventually self-sustaining – Long term sustainability - must come out of the current short term project funding cycles Sergio Bertolucci, CERN29 Large scale = long times 14th April 2010

Long term sustainability of the infrastructure Need to adapt to changing technologies – Major re-think of storage and data access – Virtualisation as a solution for job management – Complexity of the middleware compared to the actual use cases Network infrastructure – This is the most reliable service we have – Invest in networks and make full use of the distributed system (i.e. Leave data where it is)? Sergio Bertolucci, CERN30 Longer term future We have achieved what we set out to do – provide an environment for LHC computing; And we have spun-off significant general science grid infrastructures BUT: is it sustainable in the long term??? We have achieved what we set out to do – provide an environment for LHC computing; And we have spun-off significant general science grid infrastructures BUT: is it sustainable in the long term???

Grid middleware – Is still dependent upon project funding – but this is a very risky strategy now – Limited development support in EMI (for example) Must (continue) to push for mainstream, industrial solutions: – Messaging, Nagios for monitoring are good examples – Fabric and job management are good candidates for non- HEP-specific solutions Because.... Data Management is not solved – And we must invest significant effort here to improve the reliability and overall usability; must reduce complexity (e.g. SRM – functionality and implementations) – But – we are not alone – other sciences expect to have significant data volumes soon – Must take care not to have special solutions Sergio Bertolucci, CERN31 Sustainability

WLCG needs to be able to rely on strong and stable global e-science infrastructures – In Europe this means the NGIs and EGI WLCG is a very structured large user community – It can serve as a model for others – they can also learn from our mistakes CERN has connections to the other EIROs which are also large scientific communities, several of which are associated with ESFRI projects – Can play a role in bringing these to EGI CERN also supports other visible communities: – E.g. UNOSat Sergio Bertolucci, CERN32 CERN, WLCG and EGI – the future

HEP has been a leader in needing and building global collaborations in order to achieve its goals It is no longer unique – many other sciences now have similar needs – Life sciences, astrophysics, ESFRI projects – Anticipate huge data volumes – Need global collaborations There are important lessons from our experiences, – HEP was able to do this because it has a long history of global collaboration; missing from many other sciences We must also collaborate on common solutions where possible Sergio Bertolucci, CERN33 LHC is not alone

Sergio Bertolucci, CERN Summary LHC is operational and producing physics! Collaborative science on a global scale is a reality and LHC can act as a model for others 34