Presentation is loading. Please wait.

Presentation is loading. Please wait.

LHCbComputing LHCC status report. Operations June 2014 to September 2014 2 m Running jobs by activity o Montecarlo simulation continues as main activity.

Similar presentations


Presentation on theme: "LHCbComputing LHCC status report. Operations June 2014 to September 2014 2 m Running jobs by activity o Montecarlo simulation continues as main activity."— Presentation transcript:

1 LHCbComputing LHCC status report

2 Operations June 2014 to September 2014 2 m Running jobs by activity o Montecarlo simulation continues as main activity P Just in time simulation driven by analysis needs P No large simulation campaigns this quarter o User jobs o “Swimming” P On request P Complex workflow

3 Operations June 2014 to September 2014 3 m CPU usage by site o HLT farm back in business as largest site P (Previous quarter affected by power maintenance and SLC6 migration) o Otherwise business as usual

4 Job efficiency at CERN m LHCb job efficiency at CERN has been dropping throughout 2014 o Still above average though! m Only seen at CERN site (same workflows at all sites) o MC simulations not affected P Related to data access? o Investigations ongoing 4 Tier 1 CERN

5 Operations June 2014 to September 2014 m Disk usage o Smooth increase in space used by new simulation P ~50 TB / week Significant reduction in space by clearing obsolete datasets P First benefits of data popularity measurements 5

6 Disk resident data usage, last 6 months m In the last 6 months, 3.5 PB were not used, but 6.5 PB were used! LHCb week, Orsay, September 20146

7 How much was used? m Unused datasets spread in age m Needs more study why not used… m 85% of existing datasets were used in last 2 years m Still 15% could be safely retired! m 1.1 PB… LHCb week, Orsay, September 20147

8 Data processing plans, next quarter m Reprocessing of 2010 data o To provide legacy dataset with reconstruction consistent with 2011 and 2012 P Considerable work to back-port alignment and calibration o Early data (minimum bias dataset) done P Do the rest as part of Stripping21 campaign m Full restripping of 2010, 2011, 2012 data (Stripping21) o Applying latest calibrations o Intended to be legacy dataset o Used also to commission microDST for all stripping lines ahead of 2015 data-taking m Simulation of 2015 o Ready to start campaign of 2015 simulation for P Tuning of HLT bandwidth division P Systematics studies for “early measurements” 8

9 Stripping 21 status m Legacy dataset, intended to replace all previous ones o ~1500 stripping lines (c.f. 900 in Stripping 20) m Big success for microDST migration o Despite 70% more lines, S21 output bandwith is 90% of S20 m However this comes at a cost o Calculation of variables for MDST takes time P Code that was previously run in analysis, so overall gain, but slows down production P Currently ~2s/event, 3-4 times slower than S20 d Optimisation ongoing o Each stripping line has a memory footprint P Total memory usage currently too high 9

10 Stripping 21 memory usage (first test!) 10 Stripping 20 Stripping 20 + ROOT 6 Stripping 21 + ROOT 6 ROOT 6 overhead

11 Migration to ROOT 6 m Decided in May to migrate to ROOT 6, before run 2 o Using ROOT 6.00.00 as default since June P Several minor issues identified and solved promptly by development team P It was a good choice to expose users early, at a relatively quiet time, and get early feedback o Memory footprint of dictionaries is a major concern P Some improvements and workarounds available in next release P ROOT 6.02.00 scheduled for next week m Baseline for 2015 remains ROOT 6 o But dependent on resolution of memory issues, particularly in HLT environment P Benchmarking to take place with next release, early October 11

12 Software validation and deployment for 2015 m Reminder: split HLT, online calibration, single offline reconstruction pass (no reprocessing before LS2) m Goal: perform all necessary calibrations using HLT1 data and apply to both HLT2 and offline m Status: o Split HLT framework and online calibration framework implemented o Online calibration and monitoring procedures defined P Reconstruction quality demonstrated o Conditions database modifications for automatic transmission of new constants to HLT2 and offline implemented m Plans o Monthly end to end integration tests starting in October 12

13 Run 2: HLT output streams 13 m Full stream: prompt reconstruction as soon as RAW data appears offline m Parked stream: safety valve, probably not needed before 2017 m Turbo stream: no offline reconstruction, analysis objects produced in HLT o Important test for Run3

14 Turbo stream m Expect similar tracking+PID quality online and offline o Why not do “offline” event selection already online? o Suitable for analyses that don’t require offline ‘extras’ such as flavour tagging P e.g. charm spectroscopy m If event reconstructed and selected online, write MDST- like output from HLT instead of the full RAW data o Allows larger HLT rate due to x10 smaller event size o To begin with (proof of principle) write also the RAW data P But skip offline Reconstruction and Stripping steps m Basic framework already in place o Offline analysis tools running on Turbo stream output: 14

15 Run3: TurboDST? m Reminder: in Run 3, full LHCb physics potential achieved with 100kHz HLT output rate o Order of magnitude more than in Run 2 m If Run 2 Turbo stream proof of concept successful, can it be extended to a large fraction of the Run 3 HLT rate? o If only MDST content is saved (no RAW data), event size is reduced by order of magnitude m More generally, as online reconstruction approaches offline reconstruction, can we save a “DST-like” dataset out of HLT and drop the RAW data? m Currently just ideas, but would allow a 100kHz HLT output rate without order of magnitude more computing resources o Run 2 can be used as a test bed for these ideas 15

16 2015 resources m Status of 2015 pledges o Includes Russia Tier1 for the first time o For the first time in my memory, requests are fully satisfied o No worries for 2015 – thank you to all countries 16 CPU request CPU pledge Disk request Disk pledge Tape request Tape pledge kHS06 PB Tier 036 5.5 11.2 Tier 111813911.714.223.728.0 Tier 26660.51.9

17 Projections to 2016 CPU ~OK Disk ~OKTape explodes o A note on tape costs: P Costing models assume a large (~10%) disk cache in front of tape d Drives total ownership cost of tape storage P LHCb computing model does NOT require such a large buffer d ~50% of tape storage is for archival only (write once, read ~never) d Bulk recalls are during planned stripping campaigns d Can allocate temporary storage to sustain required I/O rate 17

18 Conclusions m Operations o Business as usual, no major issues o Data popularity is proving powerful tool for keeping disk usage under control m Preparations for Run 2 o Software commissioning in full swing o No concerns about resources for 2015 P Need to address tape provisioning for later years m Upgrade o Run 2 as a test bed for computing model changes for the future o Computing milestones defined: P Feb 2017: TDR P Sep 2018: Computing model for Run 3 18


Download ppt "LHCbComputing LHCC status report. Operations June 2014 to September 2014 2 m Running jobs by activity o Montecarlo simulation continues as main activity."

Similar presentations


Ads by Google