Peter Chochula.  DAQ architecture and databases  DCS architecture  Databases in ALICE DCS  Layout  Interface to external systems  Current status.

Slides:



Advertisements
Similar presentations
The Detector Control System – FERO related issues
Advertisements

CWG10 Control, Configuration and Monitoring Status and plans for Control, Configuration and Monitoring 16 December 2014 ALICE O 2 Asian Workshop
Clara Gaspar on behalf of the LHCb Collaboration, “Physics at the LHC and Beyond”, Quy Nhon, Vietnam, August 2014 Challenges and lessons learnt LHCb Operations.
André Augustinus ALICE Detector Control System  ALICE DCS is responsible for safe, stable and efficient operation of the experiment  Central monitoring.
June 23rd, 2009Inflectra Proprietary InformationPage: 1 SpiraTest/Plan/Team Deployment Considerations How to deploy for high-availability and strategies.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
Peter Chochula, January 31, 2006  Motivation for this meeting: Get together experts from different fields See what do we know See what is missing See.
Supervision of Production Computers in ALICE Peter Chochula for the ALICE DCS team.
Test Systems Software / FEE Controls Peter Chochula.
1 Data Storage MICE DAQ Workshop 10 th February 2006 Malcolm Ellis & Paul Kyberd.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Maintaining a Microsoft SQL Server 2008 Database SQLServer-Training.com.
Managing Multi-User Databases AIMS 3710 R. Nakatsu.
15 Copyright © 2005, Oracle. All rights reserved. Performing Database Backups.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
Summary DCS Workshop - L.Jirdén1 Summary of DCS Workshop 28/29 May 01 u Aim of workshop u Program u Summary of presentations u Conclusion.
JCOP Workshop September 8th 1999 H.J.Burckhart 1 ATLAS DCS Organization of Detector and Controls Architecture Connection to DAQ Front-end System Practical.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
1 Alice DAQ Configuration DB
Update on Database Issues Peter Chochula DCS Workshop, June 21, 2004 Colmar.
Peter Chochula ALICE DCS Workshop, October 6,2005 DCS Computing policies and rules.
Peter Chochula CERN/ALICE 1. 2 Experiment Size: 16 x 26 metres (some detectors placed >100m from the interaction point) Mass: 10,000,000 kg Detectors:
Peter Chochula and Svetozár Kapusta ALICE DCS Workshop, October 6,2005 DCS Databases.
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
André Augustinus 10 September 2001 DCS Architecture Issues Food for thoughts and discussion.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
Module 13 Implementing Business Continuity. Module Overview Protecting and Recovering Content Working with Backup and Restore for Disaster Recovery Implementing.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
ALICE, ATLAS, CMS & LHCb joint workshop on
P. Chochula ALICE Week Colmar, June 21, 2004 Status of FED developments.
André Augustinus 21 June 2004 DCS Workshop Detector DCS overview Status and Progress.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Peter Chochula ALICE Offline Week, October 04,2005 External access to the ALICE DCS archives.
IDE DCS development overview Ewa Stanecka, ID Week, CERN
ALICE Pixel Operational Experience R. Santoro On behalf of the ITS collaboration in the ALICE experiment at LHC.
Peter Chochula.  DCS architecture in ALICE  Databases in ALICE DCS  Layout  Interface to external systems  Current status and experience  Future.
CHAPTER 7 CLUSTERING SERVERS. CLUSTERING TYPES There are 2 types of clustering ; Server clusters Network Load Balancing (NLB) The difference between the.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Activity related CMS RPC Hyunchul Kim Korea university Hyunchul Kim Korea university 26th. Sep th Korea-CMS collaboration meeting.
The ATLAS DAQ System Online Configurations Database Service Challenge J. Almeida, M. Dobson, A. Kazarov, G. Lehmann-Miotto, J.E. Sloper, I. Soloviev and.
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
14 November 08ELACCO meeting1 Alice Detector Control System EST Fellow : Lionel Wallet, CERN Supervisor : Andre Augustinus, CERN Marie Curie Early Stage.
Summary of User Requirements for Calibration and Alignment Database Magali Gruwé CERN PH/AIP ALICE Offline Week Alignment and Calibration Workshop February.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
AB/CO Review, Interlock team, 20 th September Interlock team – the AB/CO point of view M.Zerlauth, R.Harrison Powering Interlocks A common task.
The DCS Databases Peter Chochula. 31/05/2005Peter Chochula 2 Outline PVSS basics (boring topic but useful if one wants to understand the DCS data flow)
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Summary of TPC/TRD/DCS/ECS/DAQ meeting on FERO configuration CERN,January 31 st 2006 Peter Chochula.
Database Issues Peter Chochula 7 th DCS Workshop, June 16, 2003.
André Augustinus 18 March 2002 ALICE Detector Controls Requirements.
M. Caprini IFIN-HH Bucharest DAQ Control and Monitoring - A Software Component Model.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
A Solution for Maintaining File Integrity within an Online Data Archive Dan Scholes PDS Geosciences Node Washington University 1.
«The past experience with JCOP in ALICE» «JCOP Workshop 2015» 4-5 November 2015 Alexander Kurepin.
DCS Status and Amanda News
Peter Chochula Calibration Workshop, February 23, 2005
N-Tier Architecture.
ALICE Computing Upgrade Predrag Buncic
Acutelearn Technologies Tivoli Storage Manager(TSM) Training Tivoli Storage Manager Basics: Tivoli Storage Manager Overview Tivoli Storage Manager concepts.
Juan Carlos Cabanillas Noris September 29th, 2018
DCS for LHC IF Peter Chochula 28 june 2019.
Offline framework for conditions data
Presentation transcript:

Peter Chochula

 DAQ architecture and databases  DCS architecture  Databases in ALICE DCS  Layout  Interface to external systems  Current status and experience  Future plans

 The presented slides are summary of 2 talks delivered to DATABASE FUTURES workshop  No major changes of status or planning since the june workshop

 ALICE is primary interested in ion collisions  Focus on last weeks of LHC operation in 2011 (Pb-Pb collisions)  During the year ALICE is being improved  In parallel, ALICE participates in p-p programme  So far, in 2011 ALICE delivered:  1000 hours of stable physics data taking  events collected  2.1 PB of data  5300 hours of stable cosmics datataking, calibration and technical runs  events  3.5 PB of data ▪ IONS STILL TO COME IN 2011! 4

 2011 ions approaching  First collisions already registered during LHC commissioning last weekend!

 ALICE DAQ  (slides taken from presentation of Sylvain Chapeland)

 All DB services based on MySQL  Managed by DAQ team  Tools, GUI and API developed and maintained by ALICE DAQ team  No major concerns  Satisfactory performance and scalability  Excellent stability over 3 years

 ALICE DCS is responsible for safe and correct operation of the experiment  DCS interacts with devices, configures them, monitors the operation and executes corrective actions  There are about 1200 network attached devices and ~300 directly connected devices controlled by the DCS  About parameters are actively supervised by the system  DCS interacts with many external systems  Part of the acquired data is made available to offline for analysis (conditions data)

Detector Controls System DETECTORS and DETECTOR-like systems External Services and Systems Electricity Ventilation Cooling Gas Magnets Safety Access Control LHC Configuratio n Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Infrastructure B-field Space Frame Beam Pipe Radiation Environment Alice Systems DAQ TRIGGER HLT Conditions Database ECS OFFLINE 14 SPDPHS FMDT00V00PMDMTRMCHZDCACO SDDSSDTPCTRDTOFHMP Controls Context AD0TRILHC

Detector Controls System Configuration Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Conditions Database OFFLINE CONFIGURATION DATABASE: configuration of PVSS systems device settings front-end configuration Stored mostly as code which is compiled online and sent to devices at the start of a run

Detector Controls System Configuration Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Conditions Database OFFLINE ARCHIVAL DATABASE: Parameters acquired from devices

Detector Controls System Configuration Database Archival Database Devices SCADA 1000 ins/s Up to 6GB Conditions Database OFFLINE CONDITIONS DATABASE Stores a subset of archived data Implemented at OFFLINE side Populated after each run with data acquired during the run

 All PVSS systems (sitting on ALICE network) have direct access to all archived data using PVSS built-in interface  Debugging  Interactive analysis by shift crew and experts  External and non-PVSS systems can access data only via dedicated client/server (AMANDA) suite  Protection of archive  Load balancing

 AMANDA is a ALICE-grown client- server software used to access ALICE DCS data  Client sends request for data, indicating names of the required parameters and requested time interval (without knowing the archive structure)  Server retrieves the data and sends it back to the client  Several AMANDA servers are deployed in ALICE  Multiple requests are queued in the servers and processed sequentially  In case of overload, it is enough to kill AMANDA server  AMANDA servers operate across the secured network boundary AMANDA client AMANDA server DATA REQUEST SQL req. DATA Firewall

 The main database service for the DCS is based on ORACLE  The DBA tasks are provided by DSA section of the IT-DB, based on a SLA between ALICE and IT  PRIMARY database servers and storage are located in ALICE pit  STANDBY database and tape backups are located in IT

PRIMARY DATABASE - ALICE P2 STANDBY DATABASE - IT Storage Backup SAN DB Servers ~100 DB Clients: Cofiguration Archive Offline clients Backup Streaming DATABASE – IT (limited amount of data)

 The DB is backed-up directly in ALICE site to a dedicated array  Fast recovery  Full backup  The whole DB is mirrored on STANDBY database in IT  The STANDBY database is backed up on tapes  In case of DB connectivity problems, the clients can accumulate data in local buffers and dump them to DB once the connection is restored.  Lifetime of local buffers is ~days

PRIMARY DATABASE - ALICE P2 STANDBY DATABASE - IT Storage Backup SAN DB Servers ~100 DB Clients: Cofiguration Archive Offline Backup Disaster scenario tested in 2010 All ALICE DCS redirected to standby database for several days SUCESS!!!

 Number of clients: ~100  The ALICE DCS DB is tuned and tested for:  steady insertion rate of ~1000 inserts/s  peak rate of inserts/s  Current DB size:  ~3TB  2-3 schemas/detector  Served by 6 servers and 4 SAN arrays

 ALICE DB service is in uninterrupted and stable operation since more than 3 years  Initial problems caused by instabilities of RAID arrays solved by firmware upgrades  Operational procedures fine-tuned, to match IT and ALICE requirements ▪ Updates only during LHC technical stops, etc..  The typical operational issues are caused by clients:  Misconfigured smoothing (client overload)  Missing data (stuck client, lost SOR/EOR signals) ▪ However, big improvements on stability during the last year (credits to EN-ICE guys)!

 The smooth and stable operation of the ORACLE database for ALICE DCS is a big achievement  Hereby we wish to express out thanks to the members of the IT-DB DS team for their highly professional help and approach!

 There are some additional databases deployed in ALICE DCS, but their use is very light:  MySQL – for bookkeeping on file exchange servers ▪ Zero maintenance, local backup solution  SQL Server – as a storage for system monitoring tools (MS SCOM) ▪ Used as a out-of-the box solution, but growing quickly (will need to move to a new server)  Private ORACLE server in the LAB for development

 Currently the service fulfils ALICE needs  No major architectural changes planned in the near future (before the long LHC shutdown)  HW and SW upgrades still foreseen:  A replacement of the DB service in ALICE counting room is prepared for this year winter shutdown ▪ Hardware (blade servers, SAN infrastructure and arrays) installed ▪ Software: 11G would be nice to have

 No significant increase of data volume from detectors planned before the long LHC shutdown  During the shutdown new detector modules will be added to ALICE. This might double the amount of data  New project – the ALICE-LHC interface currently store data to files (luminosities, trigger counters, etc.)  Aim to move to ORACLE  Currently estimating the load – comparable with present DCS archival

 We are getting more request for accessing the data from local analysis code  Currently were are able to satisfy the needs with AMANDA and PVSS code, but reaching the limits  Request for remote access (clients sitting on GPN or in institutes)  Read only  Mainly for debugging and monitoring purposes – latency is not an issue  Possible scenarios:  Currently we use streaming for a limited amount of data  We could use the retired service or a light DB server to contain a browsable copy of the database  11g active dataguard seems to be a very attractive solution for ALICE DCS ▪ Basic requirement – PVSS compatibility

 ALICE DAQ is based on MySQL databases managed by the DAQ team  The main DB service in ALICE DCS is based on ORACLE managed by IT  The operational experience is very positive (stability, reliability) on server side  Small issues on clients side, being constantly improved  No major modifications expected before the LHC long shutdown  Several upgrades ongoing  Again, thank to IT-DB experts for smooth operation of this critical service