Notes on offline data handling M. Moulson Frascati, 29 March 2006.

Slides:



Advertisements
Similar presentations
Status reports/Discussion items S. DellAgnello DC geometry review DC wire sag measurements P. de SimoneDC s-t relations with new sag model S. Miscetti.
Advertisements

Offline Status Report M. Moulson, 20 December 2001 Summary presentation for KLOE General Meeting Outline: Processing of 2001 data History of datarec executable.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Work finished M. Antonelli ISR and f decay simulation M. MoulsonBank-reduction code for DST’s M. Palutan et al.Trigger simulation parameters S. GiovannellaGEANFI.
Reconstruction and Analysis on Demand: A Success Story Christopher D. Jones Cornell University, USA.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
CS 333 Introduction to Operating Systems Class 18 - File System Performance Jonathan Walpole Computer Science Portland State University.
Zeus Monday Meeting, DESY, The ZEUS Event Store (ZES) How to make the most efficient data selection in ZEUS Ulrich Fricke, DESY, Hamburg.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
KLOE Offline Status Report Data Reconstruction MonteCarlo updates MonteCarlo production Computing Farm C.Bloise – May 28th 2003.
BESIII computing 王贻芳. Peak Data volume/year Peak data rate at 3000 Hz Events/year: 1*10 10 Total data of BESIII is about 2*640 TB Event size(KB)Data volume(TB)
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
Offline Discussion M. Moulson 22 October 2004 Datarec status Reprocessing plans MC status MC development plans Linux Operational issues Priorities AFS/disk.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
DBI313. MetricOLTPDWLog Read/Write mixMostly reads, smaller # of rows at a time Scan intensive, large portions of data at a time, bulk loading Mostly.
MC/Offline Planning M. Moulson Frascati, 21 March 2005 MC status Reconstruction coverage and data quality Running time estimates MC & reprocessing production.
Offline/MC status report M. Moulson 6th KLOE Physics Workshop Sabaudia, May 2006.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
KLOE Computing Update Paolo Santangelo INFN LNF KLOE General Meeting University of Rome 2, Tor Vergata 2002, December
Grand Challenge and PHENIX Report post-MDC2 studies of GC software –feasibility for day-1 expectations of data model –simple robustness tests –Comparisons.
Tutorial 9: Sequential Access Files and Printing1 Tutorial 9 Sequential Access Files and Printing.
Chapter 10 Chapter 10: Managing the Distributed File System, Disk Quotas, and Software Installation.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
CHEP 2000: 7-11 February, 2000 I. SfiligoiData Handling in KLOE 1 CHEP 2000 Data Handling in KLOE I.Sfiligoi INFN LNF, Frascati, Italy.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
DMBS Internals I. What Should a DBMS Do? Store large amounts of data Process queries efficiently Allow multiple users to access the database concurrently.
Update on MC-04(05) –New interaction region –Attenuation lengths vs endcap module numbers –Regeneration and nuclear interaction –EMC time resolution –
The PHysics Analysis SERver Project (PHASER) CHEP 2000 Padova, Italy February 7-11, 2000 M. Bowen, G. Landsberg, and R. Partridge* Brown University.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
Offline meeting  Status of MC/Datarec  Priorities for the future LNF, Sep
File Systems cs550 Operating Systems David Monismith.
Offline Status Report A. Antonelli Summary presentation for KLOE General Meeting Outline: Reprocessing status DST production Data Quality MC production.
Status report of the KLOE offline G. Venanzoni – LNF LNF Scientific Committee Frascati, 9 November 2004.
Gregory Mendell, LIGO Hanford Observatory LIGO-G WLIGO-G WLIGO-G W LIGO S5 Reduced Data Set Generation March 2007.
LHCbComputing LHCC status report. Operations June 2014 to September m Running jobs by activity o Montecarlo simulation continues as main activity.
LHCbComputing Resources requests : changes since LHCb-PUB (March 2013) m Assume no further reprocessing of Run I data o (In.
Work finished M. Antonelli ISR and f decay simulation M. MoulsonBank-reduction code for DST’s.
David Stickland CMS Core Software and Computing
Workflows and Data Management. Workflow and DM Run3 and after: conditions m LHCb major upgrade is for Run3 (2020 horizon)! o Luminosity x 5 ( )
DMBS Internals I. What Should a DBMS Do? Store large amounts of data Process queries efficiently Allow multiple users to access the database concurrently.
Offline Status Review M. Moulson, P. Valente, for the Offline Group 16 March 2001 Outline: Status update FILFO: new developments (G. Finocchiaro) Questions.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
Offline resource allocation M. Moulson, 11 February 2003 Discussion Outline: Currently mounted disk space Allocation of new disk space CPU resources.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
Status report on analysis of BR(K S  p + p - p 0 ) A. Antonelli, M. Moulson, Second KLOE Physics Workshop, Otranto, June 2002.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
1 Reminder: Why reprocess? data Reconstruction: DBV DSTs: DBV data DBV-19 for reconstruction & DSTs, Run < DBV-20 for.
DST data structures: Status report M. Moulson, CP working group 25 October 2001.
CCJ introduction RIKEN Nishina Center Kohei Shoji.
LHCb Computing activities Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Status Report on Data Reconstruction May 2002 C.Bloise Results of the study of the reconstructed events in year 2001 Data Reprocessing in Y2002 DST production.
LHCb LHCb GRID SOLUTION TM Recent and planned changes to the LHCb computing model Marco Cattaneo, Philippe Charpentier, Peter Clarke, Stefan Roiser.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
MC/OFFLINE meeting  Status  Request by Ke2 group  aob LNF
MC/OFFLINE meeting Status of data reprocessing
Compute and Storage For the Farm at Jlab
KID - KLOE Integrated Dataflow
Update on the false-dead DC wire problem
KLOE offline & computing: Status report
MC ’04-’05 objectives                                                       
Offline Report Outline: Status update DST Proposal P. de Simone
Bernd Panzer-Steindel, CERN/IT
Offline data taking and processing
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
MC/OFFLINE meeting News since the last meeting
Computing Infrastructure for DAQ, DM and SC
R. Graciani for LHCb Mumbay, Feb 2006
Presentation transcript:

Notes on offline data handling M. Moulson Frascati, 29 March 2006

Data flow in reprocessing Jobs: 1 per raw file and run (datarec_reproc_ibm.csh) Prefetch: All raw files for run Script: start_recall_files.csh, recall_one_dr.tcl Method: “recall raw (PROD areas)” List of PROD areas specified explicitly (from DB2 query) Files recalled one at a time Jobs started when files on disk Explore option of recalling all files for run at once? Input: raw files (1 per job) Script: runXreproc_ibm.csh Method: KID URL dbraw: GROUP_ID=PROD Output: datarec files (1 per stream and job) Written to /datarec and archived Advance recall to DSTPROD area: ”recall datarec dstprod”

Data flow in DST production Jobs: 1 per stream and run Disk mode (i.e. for output from reprocessing: datarec_dstfd_ibm.csh) Prefetch: None Input: All datarec files for stream and run Script: dstXprocfd_ibm.csh Method: KID URL dbdatarec GROUP_ID DSTPROD Tape mode (datarec_dst_ibm.csh) Prefetch: All datarec files for stream and run Script: start_recall_files.csh recall_one_dr.tcl Method: “recall datarec (PROD areas)” Change to DSTPROD area to avoid inconsistency? Input: datarec files Script: dstXprod_ibm.csh Method: KID URL dbdatarec GROUP_ID DSTPROD Output: DST files (1 per stream and run) Written to /datarec and archived

Data flow in MC production (1/2) Jobs: 1 per run and card type (mcprod.pl) Processes: 1 or more GEANFI processes, each followed by a datarec process 1 DST job per requested DST stream at end GEANFI output: 1 mco file per GEANFI process Written to /datarec, not archived Reconstruction prefetch: All bgg/lsb (datarec) files for run Method: “recall datarec all” Currently, prefetch all files at start of each reconstuction job Instead, prefetch once before first GEANFI process Reconstruction input: mco file Method: KID URL “ybos:” (files on /datarec) Reconstruction input: Subset of bgg/lsb files for run Method: KID URL “dbdatarec:”

Data flow in MC production (2/2) Reconstruction output: 1 mcr file per mco file (GEANFI process) Written to /datarec and archived Advance recall to DSTPROD area ”recall mc dstprod” Is this really a good idea? DSTs start right away from same directory See notes below DST input: All mcr files for job, for each requested DST stream (process) Method: KID URL dbmc DSTPROD DST output: 1 MC DST file per process Written to /datarec and archived

Data flow for standalone MC DSTs Jobs: 1 per run and card type (mcprod_dst.pl) Processes: 1 per requested DST stream Prefetch: None Input: All mcr files for run and card type Method: KID URL dbmc DSTPROD Output: 1 MC DST per process Written to /datarec and archived

Standard offline file types TypeDescriptionProduced byDSRV group kpm, ksl, rpi, rad, clb, ufo, bha Reconstructed datadatarecUSER csm, tmu, filfo, afilfo, selcos Non-reconstructed and calibration data No longer producedUSER bgg background events datarec (bgg jobs)USER? lsb background events datarecUSER? mcr Reconstructed MC events GEANFIUSER dkc, dk0, d3p, drn, drc DST filesdatarec (DST jobs)DST mkc, mk0, m3p, mrn, mrc MC DST filesdatarec (MC/MCDST jobs)DST MC files (mcr) are technically datarec streams of type ALL (stream_id = 0) DESCRIPT.STREAM_OFFLINE contains separate DSRV groups for data and MC DSRV groups shown are for data, except for mcr files, for which the MC DSRV group is shown

DSRV groups for background files All datarec types already have DSRV group dir = USER All DST types (data and MC) already have DSRV group dir = DST By default these are recalled to “DST cache” Exception is background files (bgg, lsb): 1.Change to DST group? 2.Leave as USER and recall to PROD? 3.Leave as USER and recall with kcp?

Summary of recall areas VolumeGBDSRV group recalled470DST, USER_RO recalled1470DSTPROD, PROD recalled2160DSTPROD, PROD recalled31640DST, USER_RO recalled4160USER, USER_RO recalled51640USER, USER_RO recalled63280DST, USER_RO recalled73280DST, USER_RO recalled83280DST, USER_RO DSRV groupBeforeAfter DST11950 PROD + DSTPROD USER /datarec currently 470 GB SSA disk Must add more disks to string: Access bandwidth: All MC output to /datarec 84 MB/s with 300 B80 for MC 138 MB/s if input to MCDST also from /datarec Adding disks to string helps parallelism Size: Archiver maintains /datarec at 40% full MC requires <90% full to start /datarec filled to 50% within 1 hour Archiving bandwidth: Saturated with 150 B80 for MC Must increase by system tuning: Any amount of /datarec space “immediately” filled if archiving bandwidth insufficient

Transfers to and from /datarec Transfer type Event size (KB) Rate/B80 (KB/s) Rate 300 B80 (MB/s) GEANFI output datarec input datarec output mcr archiving Recall to DSTPROD DST input DST output5103 DST archiving5103 Total (DST input DSTPROD) Total (DST input /datarec) Assumes:0.5 B80 s to fully produce 1 event, including DSTs 4 DST processes per job, zero overlap in DSTs

Recommended tape space allocation TypeCurrent (TB) Temporary allocation (TB) Final allocation (TB) raw rec DST MC MC DST Total Allocations include currently occupied space 2.MC DSTs probably appear as datarec files to archiver 3.Current library system capacity ~720 GB New cassettes will have to be ordered in future 4.Temporary allocation based on 720 GB library Assumes MC production slow 5.Final allocation assumes completion of KLOE offline program