Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.

Slides:



Advertisements
Similar presentations
Operating System.
Advertisements

SALSA HPC Group School of Informatics and Computing Indiana University.
Esma Yildirim Department of Computer Engineering Fatih University Istanbul, Turkey DATACLOUD 2013.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Control and monitoring of on-line trigger algorithms using a SCADA system Eric van Herwijnen Wednesday 15 th February 2006.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
SIDDHARTH MEHTA PURSUING MASTERS IN COMPUTER SCIENCE (FALL 2008) INTERESTS: SYSTEMS, WEB.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Event display monitoring Giuseppe Zito : Infn Bari Italy Beliy Nikita : University of Mons-Hainaut Belgium.
Data Quality Monitoring of the CMS Tracker
14th IEEE-NPSS Real Time Conference 2005, 8 June Stockholm.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
CCR GRID 2010 (Catania) Daniele Gregori, Stefano Antonelli, Donato De Girolamo, Luca dell’Agnello, Andrea Ferraro, Guido Guizzunti, Pierpaolo Ricci, Felice.
J OINT I NSTITUTE FOR N UCLEAR R ESEARCH OFF-LINE DATA PROCESSING GRID-SYSTEM MODELLING FOR NICA 1 Nechaevskiy A. Dubna, 2012.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Group Computing Strategy Introduction and BaBar Roger Barlow June 28 th 2005.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
SALSA HPC Group School of Informatics and Computing Indiana University.
Students: Anurag Anjaria, Charles Hansen, Jin Bai, Mai Kanchanabal Professors: Dr. Edward J. Delp, Dr. Yung-Hsiang Lu CAM 2 Continuous Analysis of Many.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Cracow Grid Workshop October 2009 Dipl.-Ing. (M.Sc.) Marcus Hilbrich Center for Information Services and High Performance.
CERN IT Department CH-1211 Genève 23 Switzerland t Monitoring: Tracking your tasks with Task Monitoring PAT eLearning – Module 11 Edward.
Tracker data quality monitoring based on event display M.S. Mennea – G. Zito University & INFN Bari - Italy.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
Tracker Visualization Tool: integration in ORCA Maria S. Mennea, Giuseppe Zito University & INFN Bari, Italy Tracker b-tau Cosmic Challenge preparation.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
Online Monitoring for the CDF Run II Experiment T.Arisawa, D.Hirschbuehl, K.Ikado, K.Maeshima, H.Stadie, G.Veramendi, W.Wagner, H.Wenzel, M.Worcester MAR.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
Students: Aiman Md Uslim, Jin Bai, Sam Yellin, Laolu Peters Professors: Dr. Yung-Hsiang Lu CAM 2 Continuous Analysis of Many CAMeras The Problem Currently.
Online Monitoring System at KLOE Alessandra Doria INFN - Napoli for the KLOE collaboration CHEP 2000 Padova, 7-11 February 2000 NAPOLI.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Gennaro Tortone, Sergio Fantinel – Bologna, LCG-EDT Monitoring Service DataTAG WP4 Monitoring Group DataTAG WP4 meeting Bologna –
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
14 th IEEE-NPSS Real Time Stockholm - June 9 th 2005 P. F. Zema The GNAM monitoring system and the OHP histogram presenter for ATLAS 14 th IEEE-NPSS Real.
CMS TRACKER VISUALISATION TOOLS M.S. MENNEA, a G. ZITO, a A. REGANO a AND I. OSBORNE b a Dipartimento Interateneo di fisica di Bari & INFN sezione di Bari,
VI/ CERN Dec 4 CMS Software Architecture vs Hybrid Store Vincenzo Innocente CMS Week CERN, Dec
September 2003, 7 th EDG Conference, Heidelberg – Roberta Faggian, CERN/IT CERN – European Organization for Nuclear Research The GRACE Project GRid enabled.
General Architecture of Retrieval Systems 1Adrienn Skrop.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
TIFR, Mumbai, India, Feb 13-17, GridView - A Grid Monitoring and Visualization Tool Rajesh Kalmady, Digamber Sonvane, Kislay Bhatt, Phool Chand,
G. Russo, D. Del Prete, S. Pardi Kick Off Meeting - Isola d'Elba, 2011 May 29th–June 01th A proposal for distributed computing monitoring for SuperB G.
MONITORING CMS TRACKER CONSTRUCTION AND DATA QUALITY USING A GRID/WEB SERVICE BASED ON A VISUALIZATION TOOL G. ZITO, M.S. MENNEA, A. REGANO Dipartimento.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
Cooperative Caching in Wireless P2P Networks: Design, Implementation And Evaluation.
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Grid site as a tool for data processing and data analysis
Belle II Physics Analysis Center at TIFR
ALICE Monitoring
INFN-GRID Workshop Bari, October, 26, 2004
ROBUST FACE NAME GRAPH MATCHING FOR MOVIE CHARACTER IDENTIFICATION
Ruslan Fomkin and Tore Risch Uppsala DataBase Laboratory
STORM & GPFS on Tier-2 Milan
Simulation in a Distributed Computing Environment
Presentation transcript:

Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring task was to give the integrated signal for all strips/pixels which requires building 14,248 monodimensional plots with 512/768 channels and 1392 scatter plots with up to 420x160 channels. Jobs For each of the 41 layers of the detectors 5 jobs were generated, each one processing 2000 events. Real Time of test 95 minutes with at most 35 job running in parallel. 0.5 seconds/event on average. CPU used60% of real time WN and Master Machine Pentium IV 2.4 GHz – 3.0 GHz 2GB RAM Memory ~ 300 MB used for each job The complexity of the HEP detectors for LHC (CMS Tracker has more than 50 millions of electronic channels) and the availability of the Grid computing environment require novel ways to monitor online detector performance and data quality. In the control room all raw data are accessible but the computing resources are scarce. For this reason we would like to send a sizeable amount of tracker raw data to a specialized Tier2 centre (through an high-bandwidth connection). This centre can analyze the data as they arrive, doing a quasi-online monitoring and sending back to the control room the result of this analysis. We report on a feasibility test done using a grid farm of INFN (Tier-2) in Bari. M.S. MENNEA, G. ZITO, N. DE FILIPPIS University & INFN Bari, Italy [1] CMS tracker visualization tools - Maria S.Mennea, A. Regano, G. Zito - Published in Nuclear Inst. And Methods in Physics Research A, Vol 548/3 pp , 2005 [2] Use of interactive SVG map and web services to monitor CMS tracker - G. Zito, M. S. Mennea - Proceedings of IEEE-NPSS Real Time Conference, Stockholm Sweden June 4-10, 2005 REFERENCE ARCHITECTURE AND DATA FLOW The raw data is processed by a local cluster at CERN with a few thousands CPUs connected in a hierarchical way. The final stage of this processing is the Filter Unit Farm where the raw data is made available to local monitoring programs. The same data is also sent to Tier0 and then Tier1 centers. The Tier2 will get from the nearest Tier1 center a sizeable amount of raw data through a high bandwidth connection The worker nodes (WN) available on the Tier2, as synchronously as possible with the data transfer, will start processing the data as it arrives. Each CPU will process all or a part of the tracker and only an optimal number of the events for each job. The result of each job is a root file saved on a disk local to the Tier2. In parallel with these jobs, a program will run all the time on the Master Machine waiting for the new root files to be ready Each root file is analyzed and the result is added to a kind of summary of the monitoring analysis (the In case of problems, the operator at CERN which sees the image using a web browser, can click on the module origin of the problem and get the web page containing the module's report. If a more detailed analysis of the data is requested, the operator can access also the root file available on line. This is a specialized 2D representation of the tracker, a kind of scatter plot where all modules are represented in a single screen (we imagine to disassemble the whole tracker and to assemble it again on a flat surface putting the single modules in positions which are connected to their spatial position). Using this representation in SVG format with the interactive features implemented in JavaScript, we have obtained a kind of high level user interface for tracker monitoring data visualization. This is not a static but an interactive image where the user can zoom and get more detail up to the level of microstrip in the form of a normal histogram visualized in a window nearby the main display, pick a zone and get more information on that zone, etc... SUMMARIZING THE RESULTS OF MONITORING ANALYSIS WITH A TRACKER MAP PRELIMIRARY RESULTS OF THE TEST AT TIER-2 IN BARI Limits Limits related to the test are connected to:  used 1/3 of total CPU per job because of the access of data. Rate are 100 MB/s on the server and 10 MB/s for nodes. Can be optimized using different protocols to access data: rfio, dCache. Rfio is currently used.  availability of the RB, grid overhead for the submission in a intensive use period.  retrieving output from grid job: to be replaced by the retrieve from the storage element  number of jobs running in parallel  saturation of the local area network bandwidth in the data transfer: currently it is at most 100 GB  overhead introduced by the agents for synchronous analysis after data transfer Problems Problems : failure of local (Tier2 specific) and grid services (RB. CE. SE). To be improved with the redundancy of services. TO BE OPTIMIZED Another program would start waiting for the completion of the jobs, in order to process the histograms. For this test we don't do any special check but only count the hits on a module and add the result to the number of previous hits in the same module. The result is saved periodically in the tracker map which becomes always more complete until all data is processed. Minimum delay in updating tracker map is 1 minute.  Feasibility of test  Success of preliminary test  Many parameters to be optimized: an huge improvement is expected CONCLUSIONS tracker map) which is saved on a disk area seen by server web as an SVG image. This image has entries for each of the detector modules connected through an url to a detailed report containing also the module's histograms. ABSTRACT