Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 Plans for the integration of grid tools in the CMS computing environment Claudio.

Slides:



Advertisements
Similar presentations
Claudio Grandi INFN Bologna DataTAG WP4 meeting, Bologna 14 jan 2003 CMS Grid Integration Claudio Grandi (INFN – Bologna)
Advertisements

CMS Report – GridPP Collaboration Meeting IX Peter Hobson, Brunel University4/2/2004 CMS Status Progress towards GridPP milestones Data management – the.
CMS Grid Batch Analysis Framework
Réunion DataGrid France, Lyon, fév CMS test of EDG Testbed Production MC CMS Objectifs Résultats Conclusions et perspectives C. Charlot / LLR-École.
CMS-ARDA Workshop 15/09/2003 CMS/LCG-0 architecture Many authors…
Experience with ATLAS Data Challenge Production on the U.S. Grid Testbed Kaushik De University of Texas at Arlington CHEP03 March 27, 2003.
A tool to enable CMS Distributed Analysis
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
CMS Report – GridPP Collaboration Meeting VIII Peter Hobson, Brunel University22/9/2003 CMS Applications Progress towards GridPP milestones Data management.
Dave Newbold, University of Bristol24/6/2003 CMS MC production tools A lot of work in this area recently! Context: PCP03 (100TB+) just started Short-term.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
Use of R-GMA in BOSS Henry Nebrensky (Brunel University) VRVS 26 April 2004 Some slides stolen from various talks at EDG 2 nd Review (
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
Claudio Grandi INFN Bologna ACAT'03 - KEK 3-Dec-2003 CMS Distributed Data Analysis Challenges Claudio Grandi on behalf of the CMS Collaboration.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Distribution After Release Tool Natalia Ratnikova.
CMS Report – GridPP Collaboration Meeting V Peter Hobson, Brunel University16/9/2002 CMS Status and Plans Progress towards GridPP milestones Workload management.
11 December 2000 Paolo Capiluppi - DataGrid Testbed Workshop CMS Applications Requirements DataGrid Testbed Workshop Milano, 11 December 2000 Paolo Capiluppi,
CMS Stress Test Report Marco Verlato (INFN-Padova) INFN-GRID Testbed Meeting 17 Gennaio 2003.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
29 May 2002Joint EDG/WP8-EDT/WP4 MeetingClaudio Grandi INFN Bologna LHC Experiments Grid Integration Plans C.Grandi INFN - Bologna.
13 May 2004EB/TB Middleware meeting Use of R-GMA in BOSS for CMS Peter Hobson & Henry Nebrensky Brunel University, UK Some slides stolen from various talks.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 BOSS: a tool for batch job monitoring and book-keeping Claudio Grandi (INFN Bologna)
LCG EGEE is a project funded by the European Union under contract IST LCG PEB, 7 th June 2004 Prototype Middleware Status Update Frédéric Hemmer.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
US CMS Centers & Grids – Taiwan GDB Meeting1 Introduction l US CMS is positioning itself to be able to learn, prototype and develop while providing.
Claudio Grandi INFN Bologna CMS Computing Model Evolution Claudio Grandi INFN Bologna On behalf of the CMS Collaboration.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
29 Sept 2004 CHEP04 A. Fanfani INFN Bologna 1 A. Fanfani Dept. of Physics and INFN, Bologna on behalf of the CMS Collaboration Distributed Computing Grid.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Korea Workshop May GAE CMS Analysis (Example) Michael Thomas (on behalf of the GAE group)
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
Testing the HEPCAL use cases J.J. Blaising, F. Harris, Andrea Sciabà GAG Meeting April,
David Stickland CMS Core Software and Computing
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
US ATLAS – new grid initiatives John Huth Harvard University US ATLAS Software Meeting: BNL Aug 03.
RefDB: The Reference Database for CMS Monte Carlo Production Véronique Lefébure CERN & HIP CHEP San Diego, California 25 th of March 2003.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
Computing Model José M. Hernández CIEMAT, Madrid On behalf of the CMS Collaboration XV International Conference on Computing in High Energy and Nuclear.
CMS Production Management Software Julia Andreeva CERN CHEP conference 2004.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
June 22 L. Silvestris CMS/ARDA CMS/ARDA Lucia Silvestris INFN-Bari.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
EDG Project Conference – Barcelona 13 May 2003 – n° 1 A.Fanfani INFN Bologna – CMS WP8 – Grid Planning in CMS Outline  CMS Data Challenges  CMS Production.
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
Real Time Fake Analysis at PIC
(on behalf of the POOL team)
U.S. ATLAS Grid Production Experience
Moving the LHCb Monte Carlo production system to the GRID
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
CMS Data Challenge Experience on LCG-2
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
CMS Data Challenge 2004 Claudio Grandi CMS Grid Coordinator
BOSS: the CMS interface for job summission, monitoring and bookkeeping
Scalability Tests With CMS, Boss and R-GMA
R. Graciani for LHCb Mumbay, Feb 2006
Grid Computing in CMS: Remote Analysis & MC Production
ATLAS DC2 & Continuous production
Presentation transcript:

Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 Plans for the integration of grid tools in the CMS computing environment Claudio Grandi (INFN Bologna) on behalf of the CMS-CCS group

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 2 Data Challenge 2002 on Grid Two “official” CMS productions on the grid in 2002 –CMS-EDG Stress Test on EDG testbed + CMS sites ~260K events CMKIN and CMSIM steps Top-down approach: more functionality but less robust, large manpower needed –USCMS IGT Production in the US 1M events Ntuple-only (full chain in single job) 500K up to CMSIM (two steps in single job) Bottom-up approach: less functionality but more stable, little manpower needed –See talk by P.Capiluppi

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 3 Data Challenge 2004 Next important computing milestone for CMS is the Data Challenge in 2004 (DC04) –reconstruction and analysis on CMS data sustained over one month at a rate which is the 5% of the LHC rate at full luminosity (25% of start-up luminosity) –50 millions fully digitized events needed as input –will exploit the LCG-1 resources –is a pure computing challenge! –see talk by V.Innocente for CMS data analysis

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 4 Pre-Challenge Production Simulation and digitization of 50M events (PCP04) –6 months (July to December 2003) –Transfer to CERN: ~1TB/day for 2 months (Nov.-Dec. 03) –Distributed: most of CMS Regional Centers will participate SimulationDigitizationReconstruction CPU per event (KSI2K s) Total CPU (KSI2K months) Size of output (1 event) (MB) (DST) 0.02 (AOD) Total size of sample (TB) (DST) 1 (AOD) Resource request (PIII, 1 GHz CPU, ~ 400 SI2K) 1000 for 5 months 150 for 2 months 460 (600 SI2K) for 1 month

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 5 Boundary conditions for PCP CMS persistency is changing POOL (by LCG) is replacing Objectivity/DB CMS Compiler is changing gcc is replacing Operating system is changing Red Hat 7.3 is replacing Grid middleware structure is changing EDG on top of VDT  CMS has to deal with all this while preparing for the Pre-Challenge Production!

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 6 PCP strategy PCP cannot fail (no DC04!) –basic strategy is to run on dedicated, fully controllable resources without the need of grid tools –grid-based prototypes have to be compatible with the basic non-grid environment Jobs will run in a limited-sandbox –input data local to the job –local XML POOL catalogue (prepared by the prod. tools) –output data/metadata and job monitoring data produced locally and moved to the site manager asynchronously –synchronous components optionally update central catalogues. If they fail the job will continue and the catalogues are updated asynchronously –reduce dependencies on external environment and improve robustness

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 7 Hybrid production model MCRunJob Site Manager starts an assignment RefDB Phys.Group asks for an official dataset User starts a private production Production Manager defines assignments DAG job JDL shell scripts DAGMan (MOP) Local Batch Manager EDG Scheduler Computer farm LCG-1 testbe d User’s Site Resources Chimera VDL Virtual Data Catalogue Planner

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 8 Limited-sandbox environment File transfers, if needed, are managed by external tools (EDG-JSS, additional DAG nodes, etc...) Job Wrapper (job instru- mentation) User Job Journal writer Remote updater Job input Job output Journal Metadata DB Job input Job output Journal Asynchronous updater Worker Node User’s Site

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 9 Job production Done by MCRunJob (see talk by G.Graham) –Modular: produce plug-in’s for: reading from RefDB reading from simple GUI submitting to a local resource manager submitting to DAGMan/Condor-G (MOP) submitting to the EDG scheduler producing derivations in the Chimera Virtual Data Catalogue (see talk by R.Cavanaugh) –Runs on the user (e.g. site manager) host –Defines also the sandboxes needed by the job –If needed, the specific submission plug-in takes care of: moving the sandbox files to the worker nodes preparing the XML POOL catalogue with input files information

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 10 Job Metadata management Job parameters that represent the job running status are stored in a dedicated database: –when did the job start? –is it finished? but also: –how many events did it produce so far? BOSS is a CMS-developed system that does this extracting the info from the job standard input/output/error streams –The remote updater is based on MySQL –A remote updater based on R-GMA is being developed (scalability tests being done now) for running in a grid environment –See talk by C.G.

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 11 Dataset Metadata management Dataset metadata are stored in the RefDB (see talk by V.Lefebure): –by what (logical) files is it made of? but also: –what input parameters to the simulation program? –how many events have been produced so far? Information may be updated in the RefDB in many ways: –manual Site Manager operation –automatic from the job –a remote updater similar to BOSS + R-GMA will be developed for running in a grid environment Mapping of logical names to GUID and of GUID to physical file names will be done on the grid by RLS

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 12 Other issues for PCP Software distribution and installation: –pre-installed software: rpm files installed by LCG site administrators –installed on demand (if possible): DAR files located using PACMAN or the Replica Manager –pile-up data (huge dataset!): must be pre-installed in the site (in an appropriate number of copies) to have reasonable performance on the grid considered as part of the digitization software Data transfer: –Replica Manager or direct gridFTP –MSS access using SRM under test. SE workshop…

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 13 DC04 Workflow Process data at 25 Hz (50 MB/s) at the Tier-0 –Reconstruction produces DST and AOD –AOD replicated to all Tier-1 (assume 4 centers) –DST replicated to at least one Tier-1 –Assume Digis are already replicated in at least one Tier-1 No bandwidth to transfer Digis synchronously –Archive Digis to tape library –Express lines transferred to selected Tier-1: Calibration streams, Higgs analysis stream, … Analysis & recalibration –Produce new calibration data at selected Tier-1 and update the Conditions Database –Analysis from the Tier-2 on AOD, DST, occasionally on Digis

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 14 DC04 Strategy DC04 is a computing challenge: –Run on LCG-1 (possibly integrated by CMS resources) –Use Replica Manager services to locate data –Use a Workload Management System to select resources –Use a Grid-wide monitoring system –Client-server analysis: Clarens (see talk by C.Steenberg) Data management strategy (preliminary…) –Express Lines pushed from Tier-0 to Tier-1’s –AOD, DST published by Tier-0 and pulled by Tier-1’s –Conditions DB segmented in read-only Calibration Sets Versioned Metadata stored in the RefDB Temporary solution: –need specific middleware for read-write data management

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 15 Summary Next CMS computing challenges will be done in a very dynamic environment Data Challenge 2004 will be done on LCG-1 Pre-Challenge Production already well defined –Flexible production tools may run in a local or distributed environment –Basically outside the Grid but an ideal proof of maturity for Grid-based prototypes Data Challenge architecture will be built on the experience CMS will gain during PCP

Claudio Grandi INFN Bologna March 27th 2003 CHEP'03 Conference, San Diego 16