LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3.

Slides:



Advertisements
Similar presentations
Metadata (data about data) GridPP-15, Paul Millar.
Advertisements

Database Futures Workshop CERN Michael Dahlinger, GSI
The Latest news … and Future of ATLAS Databases Elizabeth Gallas - Oxford ATLAS Software & Computing Workshop CERN November 29 to December 3, 2010.
1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
ATLAS Databases: An Overview, Athena use of Geometry/Conditions DB, and Conditions Metadata Elizabeth Gallas - Oxford ATLAS-UK Distributed Computing Tutorial.
CERN - IT Department CH-1211 Genève 23 Switzerland t Relational Databases for the LHC Computing Grid The LCG Distributed Database Deployment.
Conditions Metadata for TAGs Elizabeth Gallas, (Ryan Buckingham, Jeff Tseng) - Oxford ATLAS Software & Computing Workshop CERN – April 19-23, 2010.
29 July 2008Elizabeth Gallas1 An introduction to “TAG”s for ATLAS analysis Elizabeth Gallas Oxford Oxford ATLAS Physics Meeting Tuesday 29 July 2008.
ATLAS: Database Strategy (and Happy Thanksgiving) Elizabeth Gallas (Oxford) Distributed Database Operations Workshop CERN: November 26-27, 2009.
CERN - IT Department CH-1211 Genève 23 Switzerland t Partitioning in COOL Andrea Valassi (CERN IT-DM) R. Basset (CERN IT-DM) Distributed.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
CERN - IT Department CH-1211 Genève 23 Switzerland t The High Performance Archiver for the LHC Experiments Manuel Gonzalez Berges CERN, Geneva.
The Muon Conditions Data Management: Database Architecture and Software Infrastructure Monica Verducci University of Wuerzburg & CERN 5-9 October 2009.
Alexei Klimentov : ATLAS Computing CHEP March Prague Reprocessing LHC beam and cosmic ray data with the ATLAS distributed Production System.
Introduction: Distributed POOL File Access Elizabeth Gallas - Oxford – September 16, 2009 Offline Database Meeting.
Alignment Strategy for ATLAS: Detector Description and Database Issues
30 Jan 2009Elizabeth Gallas1 Introduction to TAGs Elizabeth Gallas Oxford ATLAS-UK Distributed Computing Tutorial January 2009.
Time and storage patterns in Conditions: old extensions and new proposals António Amorim CFNUL- FCUL - Universidade de Lisboa ● The “Extended”
ATLAS Scalability Tests of Tier-1 Database Replicas WLCG Collaboration Workshop (Tier0/Tier1/Tier2) Victoria, British Columbia, Canada September 1-2, 2007.
Software Solutions for Variable ATLAS Detector Description J. Boudreau, V. Tsulaia University of Pittsburgh R. Hawkings, A. Valassi CERN A. Schaffer LAL,
ATLAS Database Operations Invited talk at the XXI International Symposium on Nuclear Electronics & Computing Varna, Bulgaria, September 2007 Alexandre.
CHEP 2006, Mumbai13-Feb-2006 LCG Conditions Database Project COOL Development and Deployment: Status and Plans Andrea Valassi On behalf of the COOL.
2nd September Richard Hawkings / Paul Laycock Conditions data handling in FDR2c  Tag hierarchies set up (largely by Paul) and communicated in advance.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
Web application for detailed real-time database transaction monitoring for CMS condition data ICCMSE 2009 The 7th International Conference of Computational.
ALICE, ATLAS, CMS & LHCb joint workshop on
ATLAS applications and plans LCG Database Deployment and Persistency Workshop 17-Oct-2005,CERN Stefan Stonjek (Oxford), Torre Wenaus (BNL)
Metadata requirements for HEP Paul Millar. Slide 2 12 September 2007 Metadata requirements for HEP Some of the players in this game... WLCG – Umbrella.
The Persistency Patterns of Time Evolving Conditions for ATLAS and LCG António Amorim CFNUL- FCUL - Universidade de Lisboa A. António, Dinis.
1 Database mini workshop: reconstressing athena RECONSTRESSing: stress testing COOL reading of athena reconstruction clients Database mini workshop, CERN.
3rd November Richard Hawkings Luminosity, detector status and trigger - conditions database and meta-data issues  How we might apply the conditions.
CERN - IT Department CH-1211 Genève 23 Switzerland t COOL Conditions Database for the LHC Experiments Development and Deployment Status Andrea.
Conditions Metadata for TAGs Elizabeth Gallas, (Ryan Buckingham, Jeff Tseng) - Oxford ATLAS Software & Computing Workshop CERN – April 19-23, 2010.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Michele de Gruttola 2008 Report: Online to Offline tool for non event data data transferring using database.
New COOL Tag Browser Release 10 Giorgi BATIASHVILI Georgian Engineering Center 23/10/2012
Development of the CMS Databases and Interfaces for CMS Experiment: Current Status and Future Plans D.A Oleinik, A.Sh. Petrosyan, R.N.Semenov, I.A. Filozova,
Development, Deployment and Operations of ATLAS Databases XVI International Conference on Computing in High Energy and Nuclear Physics Victoria, British.
The ATLAS TAGs Database - Experiences and further developments Elisabeth Vinek, CERN & University of Vienna on behalf of the TAGs developers group.
11th November Richard Hawkings Richard Hawkings (CERN) ATLAS reconstruction jobs & conditions DB access  Conditions database basic concepts  Types.
TAGS in the Analysis Model Jack Cranshaw, Argonne National Lab September 10, 2009.
CERN IT Department CH-1211 Genève 23 Switzerland t COOL Performance Tests ATLAS Conditions Database example Romain Basset, IT-DM October.
Summary of User Requirements for Calibration and Alignment Database Magali Gruwé CERN PH/AIP ALICE Offline Week Alignment and Calibration Workshop February.
Utility of collecting metadata to manage a large scale conditions database in ATLAS Elizabeth Gallas, Solveig Albrand, Mikhail Borodin, and Andrea Formica.
The DCS Databases Peter Chochula. 31/05/2005Peter Chochula 2 Outline PVSS basics (boring topic but useful if one wants to understand the DCS data flow)
ELSSISuite Services QIZHI ZHANG Argonne National Laboratory on behalf of the TAG developers group ATLAS Software and Computing Week, 4~8 April, 2011.
Conditions Metadata for TAGs Elizabeth Gallas, (Ryan Buckingham, Jeff Tseng) - Oxford ATLAS Software & Computing Workshop CERN – April 19-23, 2010.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
M.Frank, CERN/LHCb Persistency Workshop, Dec, 2004 Distributed Databases in LHCb  Main databases in LHCb Online / Offline and their clients  The cross.
Maria del Carmen Barandela Pazos CERN CHEP 2-7 Sep 2007 Victoria LHCb Online Interface to the Conditions Database.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
ATLAS The ConditionDB is accessed by the offline reconstruction framework (ATHENA). COOLCOnditions Objects for LHC The interface is provided by COOL (COnditions.
Time critical condition data handling in the CMS experiment during the first data taking period CHEP 2010 Software Engineering, Data Stores, and Databases.
CERN - IT Department CH-1211 Genève 23 Switzerland t Persistency Framework CORAL, POOL, COOL status and plans Andrea Valassi (IT-PSS) On.
ATLAS TAGs: Tools from the ELSSI Suite Elizabeth Gallas - Oxford ATLAS-UK Distributed Computing Tutorial Edinburgh, UK – March 21-22, 2011.
Dario Barberis: ATLAS DB S&C Week – 3 December Oracle/Frontier and CondDB Consolidation Dario Barberis Genoa University/INFN.
CERN IT Department CH-1211 Genève 23 Switzerland t Load testing & benchmarks on Oracle RAC Romain Basset – IT PSS DP.
ATLAS Distributed Computing Tutorial Tags: What, Why, When, Where and How? Mike Kenyon University of Glasgow.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
L1Calo Databases ● Overview ● Trigger Configuration DB ● L1Calo OKS Database ● L1Calo COOL Database ● ACE Murrough Landon 16 June 2008.
Database Replication and Monitoring
Elizabeth Gallas - Oxford ADC Weekly September 13, 2011
Conditions Data access using FroNTier Squid cache Server
Overload of frontier lpad by MC Overlay
Data Lifecycle Review and Outlook
Elizabeth Gallas, Solveig Albrand, Mikhail Borodin, and Andrea Formica
ATLAS TAGs: Tools from the ELSSI Suite
ATLAS DC2 & Continuous production
Offline framework for conditions data
Presentation transcript:

LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3

29-Aug-2009 Elizabeth Gallas 2 Overview – Oracle usage in ATLAS Oracle used extensively at every stage of data taking and analysis Configuration PVSS – Detector Control System (DCS) Configuration & Monitoring Trigger – Trigger (Configure 3-level online event selection) OKS - Configuration databases for the TDAQ Detector Description – Geometry File and Job management Tier-0 – Initial event data processing at CERN DQ2/DDM – distributed file and dataset management Dashboard - monitor jobs and data movement on the ATLAS grid PanDa – workload management: production & distributed analysis Dataset selection catalogue AMI - dataset selection catalogue with links to other ATLAS metadata Event summary - event-level metadata TAGs – ease selection of and navigation to events of interest Conditions data (non-event data for offline analysis) Conditions Database in Oracle POOL files (referenced from the Conditions DB, stored in DDM)

29-Aug-2009 Elizabeth Gallas 3

29-Aug-2009 Elizabeth Gallas 4 Conditions Database Overview Subsystems need to store information which is needed in offline analysis which is not “event-wise”: the information represents conditions of the system for an interval ranging from very short to infinity. Volume ~ 1GB / day (online  offline) We store Conditions Data and reference to POOL files in Oracle in a generic schema design which can store/accommodate/deliver a large amount of data for a diverse set of subsystems. Relies on considerable infrastructure Based on CORAL (developed by CERN IT, ATLAS)  Restricts data types Allow extraction of ‘slices’ of Conditions into alternative DBMS Used at every stage of data taking and analysis From online calibrations and alignment … processing … more calibrations … further alignment… reprocessing … analysis … to relating recorded events with the beam conditions

29-Aug-2009 Elizabeth Gallas 5 Stages: ATLAS Reconstruction RAW data file ESD (Event Summary Data) ~500 kB/event AOD (Analysis Object Data) ~100 kB/event  TAG (not an acronym) ~ 1 kB/event (stable) Athena (framework for ATLAS reco/analysis) Input: file based events + Conditions DB via COOL COOL = the Conditions DB API, allows uniform access to ATLAS Conditions data from Athena Each stage of ATLAS processing/analysis requires Conditions DB data RAW AOD ESD TAG

29-Aug-2009 Elizabeth Gallas 6 Conditions DB terminology ATLAS Conditions Database is an Interval of Validity Database: All tables are indexed using an interval in time FOLDER: ‘think’ Table in the Conditions DB Indexed by IOV (Interval of Validity) in  time range or  run-LB range (Optionally) CHANNEL number (or name)  Useful to store many objects of identical structure  (Optionally) Version (called a COOL TAG ) Contains its ‘ PAYLOAD ’: the data (one/more columns) ‘inline’ values and/or ‘reference’ values (pointer to external) Many payload data types available – restricted by CORAL FOLDERSET: set of folders and/or foldersets arranged in hierarchical structure, names ~UNIX pathnames Subdetector Folderset_1 Folder 1 A Folder 1 B Folder_A

29-Aug-2009 Elizabeth Gallas 7 Some numbers / recent reprocessing Conditions DB data are organized in 16 database schemas: Total of 747 tables organized in 122 folders plus system tables Current volume (simulation, cosmic and commissioning data): CERN Master Online: 41 GB CERN Master Offline: 400 GB 10 Tier-1 Oracle RACs: 168 GB Volume will increase considerably with real collision data Mid-November 2009 From recent reprocessing of cosmic ray data: 35 distinct database-resident payloads from 32 bit to 16 MB Referencing 64 external POOL files in total To process a 2 GB file with 1K raw events a typical reconstruction job makes ~2K queries to read ~40 MB of database-resident data Some jobs read tens of MB extra Plus about the same volume of data is read from external POOL files

29-Aug-2009 Elizabeth Gallas 8 Challenge: Distribution of Conditions data Use cases continue to grow for distributed Processing…Calibration…Alignment…Analysis … Can happen anywhere on the ATLAS grid (worldwide) Oracle stores a huge amount of essential data Keeps all this data ‘at our fingertips’ But ATLAS has many… many… many… fingers looking for both the oldest and newest data Adding network latency  bottlenecks Solutions for distributed databases: make a ‘copy’ SQLite file: mini-Conditions DB with specific Folders, IOV range (,CoolTag) required. Considerable file management required Thought by many not to be scalable with real data Frontier: store results in a web cache (from CMS model) Located at/near the Oracle RAC