L.M.Barone – INFN Rome 16-20 October 2000 ACAT2000 - FNAL Management of Large Scale Data Productions for the CMS Experiment Presented by L.M.Barone Università.

Slides:



Advertisements
Similar presentations
L.M.Barone – INFN Rome 13 Settembre 2000 Commissione Nazionale I WP2 - Data Management L.M.Barone Università di Roma & INFN.
Advertisements

The Quantum Chromodynamics Grid James Perry, Andrew Jackson, Matthew Egbert, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Evaluation of the Globus Toolkit: Status Roberto Cucchi – INFN Cnaf Antonia Ghiselli – INFN Cnaf Giuseppe Lo Biondo – INFN Milano Francesco Prelz – INFN.
CMS HLT production using Grid tools Flavia Donno (INFN Pisa) Claudio Grandi (INFN Bologna) Ivano Lippi (INFN Padova) Francesco Prelz (INFN Milano) Andrea.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
GRID Workload Management System Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
Status of Globus activities within INFN (update) Massimo Sgaravatto INFN Padova for the INFN Globus group
Simplify your Job – Automatic Storage Management Angelo Session id:
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
INFN-Pisa Glast Database in Pisa A practical solution based on MSAccess Luca Latronico INFN Pisa.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Introduzione al Software di CMS N. Amapane. Nicola AmapaneTorino, Aprile Outline CMS Software projects The framework: overview Finding more.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
INFN-Pisa Glast Database in Pisa A practical solution based on MSAccess Luca Latronico INFN Pisa.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
MySQL and GRID Gabriele Carcassi STAR Collaboration 6 May Proposal.
Conditions DB in LHCb LCG Conditions DB Workshop 8-9 December 2003 P. Mato / CERN.
File and Object Replication in Data Grids Chin-Yi Tsai.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Efficiency of the CMS Level-1 Trigger to Selected Physics Channels by: Corey Sulkko Faculty Mentor: prof. Darin Acosta Funded by: National Science Foundation.
11 December 2000 Paolo Capiluppi - DataGrid Testbed Workshop CMS Applications Requirements DataGrid Testbed Workshop Milano, 11 December 2000 Paolo Capiluppi,
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Heterogeneous Database Replication Gianni Pucciani LCG Database Deployment and Persistency Workshop CERN October 2005 A.Domenici
4/20/02APS April Meeting1 Database Replication at Remote sites in PHENIX Indrani D. Ojha Vanderbilt University (for PHENIX Collaboration)
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
June 29, 2000DOE/NSF USCMS Computing and Software Report. HLT Studies D. Acosta1 High-Level Trigger Studies Darin Acosta University of Florida DOE/NSF.
0 Fermilab SW&C Internal Review Oct 24, 2000 David Stickland, Princeton University CMS Software and Computing Status The Functional Prototypes.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Claudio Grandi INFN-Bologna CHEP 2000Abstract B 029 Object Oriented simulation of the Level 1 Trigger system of a CMS muon chamber Claudio Grandi INFN-Bologna.
Linux Operations and Administration
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
David Adams ATLAS Datasets for the Grid and for ATLAS David Adams BNL September 24, 2003 ATLAS Software Workshop Database Session CERN.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
STAR Scheduling status Gabriele Carcassi 9 September 2002.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
Grid Workload Management (WP 1) Massimo Sgaravatto INFN Padova.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
7 Feb 2000CHEP '00. CMS/ORCA. Abstract A108. D.Stickland1 Object Reconstruction for CMS Analysis The Design Implementation and Deployment of a Functional.
Multiple Federations Issues Where to use multiple FDs How to produce FD “copies”
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
L. Perini DATAGRID WP8 Use-cases 19 Dec ATLAS short term grid use-cases The “production” activities foreseen till mid-2001 and the tools to be used.
The GridPP DIRAC project DIRAC for non-LHC communities.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
M.Frank, CERN/LHCb Persistency Workshop, Dec, 2004 Distributed Databases in LHCb  Main databases in LHCb Online / Offline and their clients  The cross.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
Workload Management Workpackage
Eleonora Luppi INFN and University of Ferrara - Italy
U.S. ATLAS Grid Production Experience
Moving the LHCb Monte Carlo production system to the GRID
ALICE FAIR Meeting KVI, 2010 Kilian Schwarz GSI.
Network Requirements Javier Orellana
US ATLAS Physics & Computing
ATLAS DC2 & Continuous production
Presentation transcript:

L.M.Barone – INFN Rome October 2000 ACAT FNAL Management of Large Scale Data Productions for the CMS Experiment Presented by L.M.Barone Università di Roma & INFN

L.M.Barone – INFN Rome October 2000 ACAT FNAL The Framework The CMS experiment is producing a large amount of MC data for the development of High Level Trigger algorithms (HLT) for fast data reduction at LHC Current production is half traditional (Pythia + CMSIM/Geant3) half OO (ORCA using Objectivity/DB)

L.M.Barone – INFN Rome October 2000 ACAT FNAL The Problem Data size ~ events, 1 MB/ev ~ 10 4 files (typically 500 evts/file) Resource dispersion many production sites CERN,FNAL,Caltech, INFN etc. Dealing with actual MC productions and not with 2005 data taking

L.M.Barone – INFN Rome October 2000 ACAT FNAL The Problem (cont’d) Data Relocation data produced in site A are stored centrally (CERN); site B may need a fraction of them; combinatorics increasing Objectivity/DB does not make life easier (but the problem would exist anyway)

L.M.Barone – INFN Rome October 2000 ACAT FNAL ORCA Production 2000 Signal Zebra files with HITS ORCA Digitization (merge signal and MB) Objectivity Database HEPEVT ntuples CMSIM HLT Algorithms New Reconstructed Objects MC Prod. ORCA Prod. HLT Grp Databases ORCA ooHit Formatter Objectivity Database MB Objectivity Database Catalog import Objectivity Database Objectivity Database ytivitcejbOesabataD Mirrored Db’s (US, Russia, Italy..)

L.M.Barone – INFN Rome October 2000 ACAT FNAL The Old Days Question: how was it done before ? A mix of ad hoc scripts/programs with a lot of manual intervention... but the problem was smaller and less dispersed

L.M.Barone – INFN Rome October 2000 ACAT FNAL Requirements for a Solution Solution must be as automatic as possible  decrease manpower Tools should be independent from data type and from site Network traffic should be optimized (or minimized ?) Users need complete information on data location

L.M.Barone – INFN Rome October 2000 ACAT FNAL Present Status Job creation is managed by a variety of scripts in different sites Job submission again goes through diverse methods, from UNIX commands to LSF or Condor File transfer has been managed up to now by Perl scripts  not generic, not site independent

L.M.Barone – INFN Rome October 2000 ACAT FNAL Present Status (cont’d) The autumn 2000 production round is a trial towards standardization  same layout (OS, installation)  same scripts (T.Wildish) for non Objy data transfer  first use of GRID tools (see talk by A.Samar)  validation procedure for production sites

L.M.Barone – INFN Rome October 2000 ACAT FNAL Collateral Activities Linux + CMS software automatic installation kit (INFN) Globus installation kit (INFN) Production monitoring tools with Web interface

L.M.Barone – INFN Rome October 2000 ACAT FNAL What is missing ? Scripts and tools are still too specific and not robust enough  need practice on this scale Information service needs a clear definition in our context and then an effective implementation (see later) File replication management is just appearing and needs careful evaluation

L.M.Barone – INFN Rome October 2000 ACAT FNAL Ideas for Replica Management A case study with Objectivity/DB (thanks to C.Grandi Bologna,INFN) –can be extended to any kind of file

L.M.Barone – INFN Rome October 2000 ACAT FNAL Cloning federations Cloned federations have a local catalog (boot file) –It is possible to manage each of them in an independent way. Some databases may be attached (or exist) only in one site –“Manual work” is needed to keep the schemas synchronized (this is not the key point today...)

L.M.Barone – INFN Rome October 2000 ACAT FNAL CERN CERNFD DB1 DB2 DB3 DBn CERN Boot RC1FD RC1 Boot RC2FD RC2 Boot Cloning federations Clone FD DB_a DB_b

L.M.Barone – INFN Rome October 2000 ACAT FNAL Productions Using a DB-id pre-allocation system it is possible to produce databases at RCs which can then be exported to other sites –A notification system is needed to inform other sites when a database is completed –This is today accomplished by GDMP using a publish-subscribe mechanism

L.M.Barone – INFN Rome October 2000 ACAT FNAL Productions When a site receives notification, it can: – ooattachdb to the remote site DB –copy the DB and ooattachdb it locally –ignore it

L.M.Barone – INFN Rome October 2000 ACAT FNAL Productions CERN CERNFD DB1 DB2 DB3 DBn CERN Boot RC1FD RC1 Boot RC2FD RC2 Boot DBn+1 DBn+m DBn+m+1 DBn+m+k

L.M.Barone – INFN Rome October 2000 ACAT FNAL Analysis In each site a complete catalog with the location of all the datasets is needed. Some DBs are local and some are remote In case more copies of a DB are available it would be nice to have in the local catalog the closest one (NWS)

L.M.Barone – INFN Rome October 2000 ACAT FNAL Information service Create an Information Service with information about all the replicas of the databases (GIS ?) In each RC there is a reference catalog which is updated taking into account the available replicas It is even possible to have a catalog created on-the-fly only for the datasets needed by a job

L.M.Barone – INFN Rome October 2000 ACAT FNAL Analysis CERN CERNFD DB1 DB2 DB3 DBn CERN Boot RC1FD RC1 Boot RC2FD RC2 Boot DBn+1 DBn+m DBn+m+1 DBn+m+k DBn+m+k DBn+m+1 DBn+m DBn+1

L.M.Barone – INFN Rome October 2000 ACAT FNAL Logical vs Physical Datasets Each dataset is composed by one or more databases –datasets are managed by application-sw Each DB is uniquely identified by a DBid –DBid assignment is a logical-db creation The physical-db is the file –zero, one or more instancies The IS manages the link between a dataset, its logical-dbs and its physical- dbs

L.M.Barone – INFN Rome October 2000 ACAT FNAL Logical vs Physical Datasets Dataset: H  2  Dataset: H  2e Hmm.1.hits.DB Hmm.2.hits.DB Hmm.3.hits.DB Hee.1.hits.DB id=12345 id=12346 id=12347 id=5678 Hee.2.hits.DB id=5679 Hee.3.hits.DB id=5680 pccms1.bo.infn.it::/data1/Hmm1.hits.DB shift23.cern.ch::/db45/Hmm1.hits.DB pccms1.bo.infn.it::/data1/Hmm2.hits.DB shift23.cern.ch::/db45/Hmm2.hits.DB shift23.cern.ch::/db45/Hmm3.hits.DB pccms5.roma1.infn.it::/data/Hee1.hits.DB shift49.cern.ch::/db123/Hee1.hits.DB pccms5.roma1.infn.it::/data/Hee2.hits.DB shift49.cern.ch::/db123/Hee2.hits.DB shift49.cern.ch::/db123/Hee3.hits.DB pccms5.roma1.infn.it::/data/Hee3.hits.DB pccms3.pd.infn.it::/data3/Hmm2.hits.DB

L.M.Barone – INFN Rome October 2000 ACAT FNAL Database creation In each production site we have: –a production federation including incomplete databases –a reference federation with only complete databases (both local and remote ones) When a DB is completed it is attached to the site reference federation The IS monitors the reference federations of all the sites and updates the database list

L.M.Barone – INFN Rome October 2000 ACAT FNAL RC1Prod DB4 DB4 RC1Ref pc.rc1.net CERN CERNFD DB1 DB2 DB3 DB DB1.DB shift.cern.ch::/shift/data 0002 DB2.DB shift.cern.ch::/shift/data 0003 DB3.DB shift.cern.ch::/shift/data 0004 DB4.DB pc.rc1.net::/pc/data shift.cern.ch::/shift/data 0005 Database creation DB5 DB5 DB5 DB DB1.DB shift.cern.ch::/shift/data 0002 DB2.DB shift.cern.ch::/shift/data 0003 DB3.DB shift.cern.ch::/shift/data 0004 DB4.DB pc.rc1.net::/pc/data shift.cern.ch::/shift/data 0005 DB5.DB pc.rc1.net::/pc/data 0001 DB1.DB shift.cern.ch::/shift/data 0002 DB2.DB shift.cern.ch::/shift/data 0003 DB3.DB shift.cern.ch::/shift/data 0004 DB4.DB pc.rc1.net::/pc/data shift.cern.ch::/shift/data 0005 DB5.db pc.rc1.net::/ps.data shift.cern.ch::/shift/data shift.cern.ch

L.M.Barone – INFN Rome October 2000 ACAT FNAL Replica Management In case of multiple copies of the same DB each site may choose which copy to use: –it should be possible to update the reference federation at given times –it should be possible to create on-the-fly a mini-catalog only with information about the datasets requested by a job this kind of operation is managed by application- sw (e.g. ORCA)

L.M.Barone – INFN Rome October 2000 ACAT FNAL 0001 DB1.DB shift.cern.ch::/shift/data pc1.bo.infn.it::/data 0002 DB2.DB shift.cern.ch::/shift/data 0003 DB3.DB shift.cern.ch::/shift/data 0001 DB1.DB shift.cern.ch::/shift/data pc1.bo.infn.it::/data 0002 DB2.DB shift.cern.ch::/shift/data pc1.bo.infn.it::/data 0003 DB3.DB shift.cern.ch::/shift/data Replica Management CERN CERNFD DB1 DB2 DB3 BORef shift.cern.ch pc1.bo.infn.it PDRef pc1.pd.infn.it DB1 DB2

L.M.Barone – INFN Rome October 2000 ACAT FNAL Summary of the Case Study Basic functionalities of a Replica Manager for production are already implemented in GDMP The use of an Information Server would allow easy synchronization of federations and optimized data access during analysis The same functionalities offered by the Objectivity/DB catalog may be implemented for other kind of files

L.M.Barone – INFN Rome October 2000 ACAT FNAL Conclusions (?) Globus and the various GRID projects try to address the issue of Large Scale distributed data access Their effectiveness is still to be proven The problem again is not the software, it is the organization