November 11-17 SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.

Slides:



Advertisements
Similar presentations
1 14 Feb 2007 CMS Italia – Napoli A. Fanfani Univ. Bologna A. Fanfani University of Bologna MC Production System & DM catalogue.
Advertisements

Workload management Owen Maroney, Imperial College London (with a little help from David Colling)
INFSO-RI Enabling Grids for E-sciencE Workload Management System and Job Description Language.
1 CRAB Tutorial 19/02/2009 CERN F.Fanzago CRAB tutorial 19/02/2009 Marco Calloni CERN – Milano Bicocca Federica Fanzago INFN Padova.
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
Réunion DataGrid France, Lyon, fév CMS test of EDG Testbed Production MC CMS Objectifs Résultats Conclusions et perspectives C. Charlot / LLR-École.
Workload Management meeting 07/10/2004 Federica Fanzago INFN Padova Grape for analysis M.Corvo, F.Fanzago, N.Smirnov INFN Padova.
INFSO-RI Enabling Grids for E-sciencE EGEE Middleware The Resource Broker EGEE project members.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
A tool to enable CMS Distributed Analysis
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Physicists's experience of the EGEE/LCG infrastructure usage for CMS jobs submission Natalia Ilina (ITEP Moscow) NEC’2007.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
LcgCAF:CDF submission portal to LCG Federica Fanzago for CDF-Italian Computing Group Gabriele Compostella, Francesco Delli Paoli, Donatella Lucchesi, Daniel.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
RISICO on the GRID architecture First implementation Mirko D'Andrea, Stefano Dal Pra.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Interactive Job Monitor: CafMon kill CafMon tail CafMon dir CafMon log CafMon top CafMon ps LcgCAF: CDF submission portal to LCG resources Francesco Delli.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Grid infrastructure analysis with a simple flow model Andrey Demichev, Alexander Kryukov, Lev Shamardin, Grigory Shpiz Scobeltsyn Institute of Nuclear.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
DBS/DLS Data Management and Discovery Lee Lueking 3 December, 2006 Asia and EU-Grid Workshop 1-4 December, 2006.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
Daniele Spiga PerugiaCMS Italia 14 Feb ’07 Napoli1 CRAB status and next evolution Daniele Spiga University & INFN Perugia On behalf of CRAB Team.
INFSO-RI Enabling Grids for E-sciencE Using of GANGA interface for Athena applications A. Zalite / PNPI.
1. 2 Overview Extremely short summary of the physical part of the conference (I am not a physicist, will try my best) Overview of the Grid session focused.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Joint Institute for Nuclear Research Synthesis of the simulation and monitoring processes for the data storage and big data processing development in physical.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Enabling Grids for E-sciencE Work Load Management & Simple Job Submission Practical Shu-Ting Liao APROC, ASGC EGEE Tutorial.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
DIRAC: Workload Management System Garonne Vincent, Tsaregorodtsev Andrei, Centre de Physique des Particules de Marseille Stockes-rees Ian, University of.
Real Time Fake Analysis at PIC
Eleonora Luppi INFN and University of Ferrara - Italy
Introduction to Grid Technology
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
LCG middleware and LHC experiments ARDA project
N. De Filippis - LLR-Ecole Polytechnique
Grid Computing in CMS: Remote Analysis & MC Production
Presentation transcript:

November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team

November SC06 Tampa F.Fanzago 2 CMS (Compact Muon Solenoid) is one of the four particle physics experiment that will collect data at LHC (Large Hadron Collider) starting in 2007 at CERN. CMS will produce a big quantity of data Data should be stored and made available for analysis to world-wide distributed physicists. CMS overview ~2 PB events/year (startup luminosity 2x10 33 cm -2 s -1 ) All events will be stored into files –O(10^6) files/year Files will be grouped in Fileblocks –O(10^3) Fileblocks/year Fileblocks will be grouped in Datasets –O(10^3) Datasets (total after 10 years of CMS) – TB

November SC06 Tampa F.Fanzago 3 Why the grid… ISSUES: How to manage and where to store this huge quantity of data? How to assure data access to physicists of CMS collaboration? How to have enough computing power for processing and data analysis? How to ensure resources and data availability? How to define local and global policy about data access and resources? SOLUTION: CMS will use a distributed architecture based on grid infrastructure to ensure remote resources availability and to assure remote data access to authorized user (belonging to CMS Virtual Organization). The grid infrastructure guarantees also enough computing power for simulation, processing and analysis data.

November SC06 Tampa F.Fanzago 4 CMS computing model Online system Tier 0 Tier 1 Tier 2 Tier 3 Offline farm. Tier2 Center InstituteB InstituteA... workstation Italy Regional Center Fermilab Regional Center France Regional Center recorded data The CMS offline computing system is arranged in four Tiers which are geographically distributed Remote data accessible via grid CERN Computer center

November SC06 Tampa F.Fanzago 5 During data acquisition data from detector which got over different trigger level will be sent, stored and first step reconstructed at Tier-0. Then they will be distributed over some Tiers depending on the kind of physics data Until real data are not available, the CMS community needs simulated data to study the detector response, the foreseen physics interaction and to get experience with management and analysis of data. So a large number of simulated data are produced and distributed among computing centers. Data distribution…

November SC06 Tampa F.Fanzago 6 grid middleware… Resource Broker (RB) Workload Management System SE UI Job submission tools UI Job submission tools Data location system Data location system Information Service collector Information Service collector Query for data Query for matchmaking CE Main LCG middleware components: Virtual Organizations (CMS...) Resource Broker (RB) Replica Catalog (LFC) Computing Element (CE) Storage Element (SE) Worker node (WN) User Interface (UI) LCG middleware: Tools for accessing distributed data and resources are provided by the World LHC Computing Grid (WLCG) that takes care about different grid flavours as LCG/gLite in Europe and OSG in the US.

November SC06 Tampa F.Fanzago 7 Analysis in a local environment… But now data and resources are distributed User writes his own analysis code and configuration parameter card –Starting from CMS specific analysis software –Builds executable and libraries He apply the code to a given amount of events, whose location is known, splitting the load over many jobs –But generally he is allowed to access only local data He writes wrapper scripts and uses a local batch system to exploit all the computing power –Comfortable until data you’re looking for are sitting just by your side Then he submits all by hand and checks the status and overall progress Finally collects all output files and store them somewhere

November SC06 Tampa F.Fanzago 8 …and in a distributed environment The distributed analysis is a more complex computing task because it assume to know: which data are available where data are stored and how to access them which resources are available and are able to comply with analysis requirements grid and CMS infrastructure details Users don't want deal with these kind of problem They want to analyze data in “a simple way” as in local environment

November SC06 Tampa F.Fanzago 9 Distributed analysis chain To allow distributed analysis the CMS collaboration is developing some tools interfaced with available grid services, that include: Installation of CMS software via grid on remote resources Data transfer service: to move and manage a large flow of data among tiers Data validation system: to ensure data consistency Data location system: catalogues to keep track of data available in each site and to allow data discovery –Dataset Bookkeeping System. It knows which data exist and contains CMS specific description of event data –Data Location Service. It knows where data are stored. Mapping between file-blocks and SE –Local file catalog: physical location of local data on remote SE CRAB: Cms Remote Analysis Builder...

November SC06 Tampa F.Fanzago 10 CRAB CMS Remote Analysis Builder CRAB is a user-friendly tool whose aim is to simplify the work of users with no knowledge of grid infrastructure to create, submit and manage job analysis into grid environments. –written in python and installed on UI (grid user access point) Users have to develop their analysis code in a interactive environment decide which data to analyze and how to manage jobs output Data discovery on remote resources, resources availability, status monitoring and output retrieval of submitted jobs are fully handled by CRAB UI with CRAB User analysis code as in local environment data in remote and distributed sites

November SC06 Tampa F.Fanzago 11 CRAB workflow dataset n.of events user code Data Location System SE Local File Catalog Data Bookkeeping System Job jdl tgz sh WMS Job CE WN... Job output CRAB UI SEs list data Main CRAB functionalities: Input data discovery: the list of sites (SEs name) where data are stored, querying “data location system” (DBS-DLS) Packaging of user code: creation of a tgz archive with user code and parameters Job creation: Wrapper of user code executable to run on WN (sh script) Jdl file: SE name as requirement to drive resources matchmaking Job splitting according to user request Job submission to the grid Job status monitoring and output retrieval Handling of user output: copy to UI or to a generic Storage Element Job resubmission in case of failure

November SC06 Tampa F.Fanzago 12 CRAB usage (1) Job submitted from worldwide distributed UI The total number of jobs submitted to the grid using CRAB is more than 1 million by users. SC03 PTDR SC04 More or less 1500 jobs are submitted each day.

November SC06 Tampa F.Fanzago 13 CRAB usage (2) ~75% job success rate (success means jobs arrive to remote sites and produce outputs) ~25% aborts due to site setup problem or grid services failure. physicists are using CRAB to analyze remote data stored in LCG and OSG sites. ~7000 Datasets available for O(10^8) total events, full MC production

November SC06 Tampa F.Fanzago 14 CRAB tool is used to analyze remote data and also to continuously test CMS Tiers to prove the whole infrastructure robustness CRAB proves a CMS user with no knowledge about grid infrastructure are able to use grid services CRAB demonstrates distributed analysis works in a distributed environment. The future code development will be related to split CRAB in a client-server system with the aim to minimize the user effort to manage analysis jobs and obtain their results. Conclusion