Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Clara Gaspar on behalf of the LHCb Collaboration, “Physics at the LHC and Beyond”, Quy Nhon, Vietnam, August 2014 Challenges and lessons learnt LHCb Operations.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
LHCb Quarterly Report October Core Software (Gaudi) m Stable version was ready for 2008 data taking o Gaudi based on latest LCG 55a o Applications.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
LHCb computing in Russia Ivan Korolko (ITEP Moscow) Russia-CERN JWGC, October 2005.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
LHCb The LHCb Data Management System Philippe Charpentier CERN On behalf of the LHCb Collaboration.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
WLCG and the India-CERN Collaboration David Collados CERN - Information technology 27 February 2014.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
Status report of the KLOE offline G. Venanzoni – LNF LNF Scientific Committee Frascati, 9 November 2004.
LHCbComputing LHCC status report. Operations June 2014 to September m Running jobs by activity o Montecarlo simulation continues as main activity.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
LHCb Readiness for Run WLCG Workshop Okinawa
OPERATIONS REPORT JUNE – SEPTEMBER 2015 Stefan Roiser CERN.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
tons, 150 million sensors generating data 40 millions times per second producing 1 petabyte per second The ATLAS experiment.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
LHCbComputing LHCb computing model in Run1 & Run2 Concezio Bozzi Bologna, Feb 19 th 2015.
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
Predrag Buncic CERN ALICE Status Report LHCC Referee Meeting 24/05/2015.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
THE ATLAS COMPUTING MODEL Sahal Yacoob UKZN On behalf of the ATLAS collaboration.
First collisions in LHC
LHCb distributed computing during the LHC Runs 1,2 and 3
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Grid site as a tool for data processing and data analysis
Predrag Buncic ALICE Status Report LHCC Referee Meeting CERN
evoluzione modello per Run3 LHC
LHC experiments Requirements and Concepts ALICE
Data Challenge with the Grid in ATLAS
Beam Gas Vertex – Beam monitor
Offline data taking and processing
LHCb Software & Computing Status
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
The LHC Computing Grid Visit of Her Royal Highness
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Project Status Report Computing Resource Review Board Ian Bird
Issues with Simulating High Luminosities in ATLAS
Simulation use cases for T2 in ALICE
ALICE Computing Model in Run3
US CMS Testbed.
ALICE Computing Upgrade Predrag Buncic
CPU efficiency Since May CMS O+C has launched a dedicated task force to investigate CMS CPU efficiency We feel the focus is on CPU efficiency is because.
R. Graciani for LHCb Mumbay, Feb 2006
LHCb Computing Philippe Charpentier CERN
DØ MC and Data Processing on the Grid
ATLAS DC2 & Continuous production
LHCb thinking on Regional Centres and Related activities (GRIDs)
The LHC Computing Grid Visit of Professor Andreas Demetriou
Study of Top properties at LHC
The LHCb Computing Data Challenge DC06
Presentation transcript:

Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group

Guidelines for the Computing Model Small processing time, but high trigger rate 24 kHS06 required for reconstruction Typically 2000 CPU slots Tier0 could not provide the necessary CPU power Use Tier1s as well for reconstruction (first pass) Most problems for analysis jobs are related to Data Management SE accessibility, scalability, reliability… Restrict the number of sites with data access Use Tier1s for analysis High requirements on simulated data Background identification, efficiency estimation for signal Typically 360 HS06.s per event Use all possible non-Tier1 resources for simulation Philippe.Charpentier@cern.ch

The LHCb Computing Model Philippe.Charpentier@cern.ch

LHCb real data in 2010 LHC started with very low luminosity Very few colliding bunches Not worth for rare b-physics decays Minimum bias trigger for 2 months Introduce tighter triggers when luminosity increases LHC change of strategy for higher luminosity Large number of protons per bunch Small squeezing Still low number of bunches (16, 25, 48, increasing since September, up to 400 bunches) Consequence: larger number of collisions per crossing µ=1 to 2.3 !!! Much higher pile-up (1.6 to 2.3 collisions per trigger) Effects on Computing Larger events More complex events to reconstruct Larger pre-selection retention Philippe.Charpentier@cern.ch

Adaptability of the Computing Model Needs to be reactive to continuously changing conditions First months: minimum bias data No preselection Reconstruction creating DSTs for all events As of July: large µ data Event size 50 to 60 kB Twice more than design Reconstruction time Quadratic with event size 4 times more than design Stripping time and memory Large combinatorics for pre-selection Stripping time exponential with event size Algorithms tuned for µ=0.4 were taking up to 60 HS06.s Twice the reconstruction time! Memory consumption up to 3 GB (nominally 1.5 GB) Philippe.Charpentier@cern.ch

New Computing Conditions Both event size and CPU time rise with µ Compatible with expectations at µ=0.4 Brighter colors Put a reference Philippe.Charpentier@cern.ch

Weaknesses of the Grid Workload Management System Mitigated by usage of pilot jobs (DIRAC) Workload optimisation using generic pilot jobs Run multiple payloads (e.g. production + analysis) Data Management System Data access by protocol unreliable for long jobs Errors when opening files (servers overloaded) Connections broken when job lasts hours Use as few files as possible, i.e. as large as possible Requires merging of output files (DSTs) Keep runs (1 hour data taking) as granularity of datasets Mitigated by local copy of input data Standard procedure for reconstruction-stripping jobs and merging Not possible for analysis jobs though… Skip a bit Philippe.Charpentier@cern.ch

Data collection 42.2 pb-1 delivered 37.7 pb-1 recorded 89.3% data taking efficiency Most data collected after 15 September Philippe.Charpentier@cern.ch

RAW data distribution 65.7 TB of physics RAW data collected Slightly more transferred to Castor Calibration and test data Not distributed to Tier1s Distributed immediately to Tier1s A full run (1 hour) goes to a single Tier1 RAW data share according to CPU pledges of Tier1s When a Tier1 is unavailable, share temporarily set to 0 Philippe.Charpentier@cern.ch

Global Grid usage 121 sites used Since April 2010: 22 countries Simulation: 72% Mainly since December Analysis: 17% Reconstruction: 11% Put labels Philippe.Charpentier@cern.ch

Reconstruction jobs Continuous processing From time to time reprocess all existing data when changes are worth it Major one in August Remove prodsand enlarge plot Philippe.Charpentier@cern.ch

Analysis jobs Over 250 users used the Grid for analysis Only 2% of analysis at Tier2s (toy MC, private small simulations) No a-priori assignment of site: share by availability of resources and data Enlarge plot, reduce text Philippe.Charpentier@cern.ch

Job success Overall 81% successful jobs Main cause of failures (15%): job exceeding CPU time limit Infinite loop in Geant4 on 64-bit Large µ Jobs eventually all completed after several retries! Also few user jobs 4% data access problem in application Simulation CPU limit as well Philippe.Charpentier@cern.ch

Winter re-processing Re-processing of all 2010 data New calibration New pre-selection (stripping) algorithms Philippe.Charpentier@cern.ch

Winter Simulation Campaign Large campaign of simulation for extracting physics from 2010 data Uses the beam and detector conditions of 2010 Steady pace of 32,000 jobs per day (about 5 million events) Already half of the highest priority events have been simulated Philippe.Charpentier@cern.ch

Conclusions The LHCb Computing Model looks global sound However the new LHC running conditions imply some changes to the offline reconstruction and analysis conditions During 2010, several iterations were needed in order to adapt to these conditions The full reprocessing of 2010 data will take place starting in November 2010 Increase in CPU requirements and disk space will have to be watched carefully in order to match the pledges Usage or resources beyond Tier1s for reconstruction and analysis are being investigated Philippe.Charpentier@cern.ch