The ALICE data quality monitoring Barthélémy von Haller CERN PH/AID For the ALICE Collaboration.

Slides:



Advertisements
Similar presentations
TeraGrid Deployment Test of Grid Software JP Navarro TeraGrid Software Integration University of Chicago OGF 21 October 19, 2007.
Advertisements

IEEE NSS 2003 Performance of the Relational Grid Monitoring Architecture (R-GMA) CMS data challenges. The nature of the problem. What is GMA ? And what.
Data Quality Monitoring workshop. What this presentation is not What it is and how it is organized  Definition of DQM  Overview of systems and.
Kondo GNANVO Florida Institute of Technology, Melbourne FL.
CWG10 Control, Configuration and Monitoring Status and plans for Control, Configuration and Monitoring 16 December 2014 ALICE O 2 Asian Workshop
GNAM and OHP: Monitoring Tools for the ATLAS Experiment at LHC GNAM and OHP: Monitoring Tools for the ATLAS Experiment at LHC M. Della Pietra, P. Adragna,
DQM news Technical side. Web tools  Advantages  Can be ran out of P2 with access rights  Centrally maintained -> can’t be altered on machines  Modern.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/27 A Control Software for the ALICE High Level Trigger Timm.
A Database Visualization Tool for ATLAS Monitoring Objects A Database Visualization Tool for ATLAS Monitoring Objects J. Batista, A. Amorim, M. Brandão,
Trigger and online software Simon George & Reiner Hauser T/DAQ Phase 1 IDR.
Control and monitoring of on-line trigger algorithms using a SCADA system Eric van Herwijnen Wednesday 15 th February 2006.
First year experience with the ATLAS online monitoring framework Alina Corso-Radu University of California Irvine on behalf of ATLAS TDAQ Collaboration.
Test results Test definition (1) Istituto Nazionale di Fisica Nucleare, Sezione di Roma; (2) Istituto Nazionale di Fisica Nucleare, Sezione di Bologna.
Quality Control B. von Haller 8th June 2015 CERN.
Architecture Of ASP.NET. What is ASP?  Server-side scripting technology.  Files containing HTML and scripting code.  Access via HTTP requests.  Scripting.
Central Online Grading System COGS Dec15-21 dec1521.sd.ece.iastate.edu.
Tools and software process for the FLP prototype B. von Haller 9. June 2015 CERN.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
ALICE Electronic Logbook MEST-CT Vasco Barroso PH/AID.
The Application of DAQ-Middleware to the J-PARC E16 Experiment E Hamada 1, M Ikeno 1, D Kawama 2, Y Morino 1, W Nakai 3, 2, Y Obara 3, K Ozawa 1, H Sendai.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Online Monitoring with MonALISA Dan Protopopescu Glasgow, UK Dan Protopopescu Glasgow, UK.
J.T Moscicki CERN LCG - Software Process & Infrastructure1 SPI Software Process & Infrastructure for LCG Software Packaging and Distribution LCG Application.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
Ramiro Voicu December Design Considerations  Act as a true dynamic service and provide the necessary functionally to be used by any other services.
F. Brasolin / A. De Salvo – The ATLAS benchmark suite – May, Benchmarking ATLAS applications Franco Brasolin - INFN Bologna - Alessandro.
1 Alice DAQ Configuration DB
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Experiences, limitations and suggested improvements The ALICE DQM Software and ROOT ROOT Users Workshop Barthelemy von Haller & Adriana Telesca for the.
Contents 1.Introduction, architecture 2.Live demonstration 3.Extensibility.
AMORE Quality Assurance Integration Filimon Roukoutakis CERN Physics Department ALICE DAQ ALICE Offline Week CERN, 11/10/2007.
May PEM status report. O.Bärring 1 PEM status report Large-Scale Cluster Computing Workshop FNAL, May Olof Bärring, CERN.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
Personal Computer - Stand- Alone Database  Database (or files) reside on a PC - on the hard disk.  Applications run on the same PC and directly access.
Control in ATLAS TDAQ Dietrich Liko on behalf of the ATLAS TDAQ Group.
1 / 22 AliRoot and AliEn Build Integration and Testing System.
Roberto Divià, CERN/ALICE 1 CHEP 2009, Prague, March 2009 The ALICE Online Data Storage System Roberto Divià (CERN), Ulrich Fuchs (CERN), Irina Makhlyueva.
ALICE, ATLAS, CMS & LHCb joint workshop on
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
AgentsAnd Daemons Automating Data Quality Monitoring Operations Agents And Daemons Automating Data Quality Monitoring Operations Since 2009 when the LHC.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
1 “Steering the ATLAS High Level Trigger” COMUNE, G. (Michigan State University ) GEORGE, S. (Royal Holloway, University of London) HALLER, J. (CERN) MORETTINI,
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
EMu Interface and the Web Clear identification of web fields for users and administrators Visual identifier of the web presentations in EMu, ie Collection.
News on GEM Readout with the SRS, DATE & AMORE
1 ITS Quality Assurance (& DQM) P. Cerello, P. Christakoglou, W. Ferrarese, M. Nicassio, M. Siciliano ALICE OFFLINE WEEK – April 2008.
September 2007CHEP 07 Conference 1 A software framework for Data Quality Monitoring in ATLAS S.Kolos, A.Corso-Radu University of California, Irvine, M.Hauschild.
April 2003 Iosif Legrand MONitoring Agents using a Large Integrated Services Architecture Iosif Legrand California Institute of Technology.
G.Govi CERN/IT-DB 1 September 26, 2003 POOL Integration, Testing and Release Procedure Integration  Packages structure  External dependencies  Configuration.
By Douglas Copas and Mark Perry.  Currently available small business based commercial inventory management systems are either prohibitively expensive.
RunControl status update Nicolas Lurkin School of Physics and Astronomy, University of Birmingham NA62 TDAQ Meeting – CERN, 14/10/2015.
The ATLAS DAQ System Online Configurations Database Service Challenge J. Almeida, M. Dobson, A. Kazarov, G. Lehmann-Miotto, J.E. Sloper, I. Soloviev and.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
The DCS Databases Peter Chochula. 31/05/2005Peter Chochula 2 Outline PVSS basics (boring topic but useful if one wants to understand the DCS data flow)
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
R.Divià, CERN/ALICE 1 ALICE off-line week, CERN, 9 September 2002 DAQ-HLT software interface.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
M. Caprini IFIN-HH Bucharest DAQ Control and Monitoring - A Software Component Model.
DGAS Distributed Grid Accounting System INFN Workshop /05/1009, Palau Giuseppe Patania Andrea Guarise 6/18/20161.
CERN IT Department CH-1211 Genève 23 Switzerland t Load testing & benchmarks on Oracle RAC Romain Basset – IT PSS DP.
Barthélémy von Haller CERN PH/AID For the ALICE Collaboration The ALICE data quality monitoring system.
DAQ thoughts about upgrade 11/07/2012
QC-specific database(s) vs aggregated data database(s) Outline
Slow Control and Run Initialization Byte-wise Environment
Slow Control and Run Initialization Byte-wise Environment
Controlling a large CPU farm using industrial tools
Support for ”interactive batch”
Module 01 ETICS Overview ETICS Online Tutorials
Offline framework for conditions data
Presentation transcript:

The ALICE data quality monitoring Barthélémy von Haller CERN PH/AID For the ALICE Collaboration

The ALICE experiment LHC : Large Hadron Collider ALICE : A Large Ion Collider Experiment – 18 detectors – Bandwidth to mass storage : 1.25 GB/s – Event size : 86.5 MB – Trigger rate : 10 KHz 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID1/15

Data Quality Monitoring 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID2/24 Online feedback on the quality of data Avoid taking and recording low-quality data Identify and solve problem(s) early Data Quality Monitoring (DQM) involves -Online gathering of data -Analysis by user-defined algorithm -Storage of monitoring data -Visualization

Data-Acquisition architecture 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID3/24 DA DQM Sub-event

The AMORE framework AMORE : Automatic MOnitoring Environment A DQM framework for the ALICE experiment 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID4/24

Publisher – Subscriber paradigm Database used for the data pool Notification with DIM (Distributed Information Management System) 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID5/24 Design & Architecture

Published objects are encapsulated into « MonitorObject » structure Plugin architecture using ROOT reflection – Modules are dynamic libraries loaded at runtime 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID6/24 Design & Architecture

14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID7/24

The Pool 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID8/24 Current implementation based on a database MySQL : reliable, performant, open-source

Archiving Short-term history : First-In First-Out (FIFO) Long-term archives : At end of run, regular intervals, and users’ request 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID9/24 Agent Latest value GUI Publish Access X recent values FIFO Temporary and permanent archive Archive triggers : Start and end of run, regular time interval, at shifter’s request

Subscriber & User Interface Generic GUI – Display any object of any running agent – Possibility of handling automatically the layout – Layout can be pretty complex and saved for future reuse – Fit the basic needs of the users to check what is published by the agents For more complex needs, users can develop their own GUI 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID10/24

The generic GUI 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID11/24 Agent Agents Monitor ObjectsSub-directories

The generic GUI 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID12/24 Save Load

Custom gui 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID13

Packaging & validation Subversion repositories GNU Autotools Distributed as RPM (1+12 packages) Strict release procedure – Build and validate the module on a test machine in a clean and controlled environment Nightly build – Identify broken code (wrong results, unable to compile) 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID14/24

Performance & benchmark 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID15/24 Online environment and heavy calculation  ensure performance and scalability To identify and handle performance issues we need : – Metrics – Statistics – Reproducible tests

Performance & benchmark Same procedure and environment as for the validation of modules – Estimation of needs for each detector – Identification of variations over time – Comparisons of machines, compilers and architectures 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID16/24

Performances & benchmark 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID17/24 Current DQM nodes : Intel(R) Xeon(R) CPU Latest generation of intel processor : Intel(R) Core(TM) i7 CPU

Database benchmark All data transit through the pool  critical part of the system Test of extreme and standard use cases Several improvements made : – Concatenate queries and insertions – MySQL engine : MyISAM vs InnoDB 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID18/24

InnoDB vs MyISAM 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID19/24

Status In production since last summer, used during commissioning and first beam 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID20/24

14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID21/24

Status In production since last summer, used during commissioning and first beam New features are regularly added, usually at users’ request 18 modules under development 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID22/24

Plans Access to monitor objects through the web via the ALICE electronic LogBook Fully automatize the process : comparisons to reference data, identification of problems, notification, actions taken Add features to take full advantage of multi-cores architecture 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID23/24

Conclusion AMORE has been in production for almost a year Increasing number of detector agents Proved to be very useful during commissioning and first beam period Capable of handling large number of agents, clients and objects  Ready for the LHC restart ! 14/05/2009 – IEEE RT2009Barthélémy von Haller - CERN PH/AID24/24