LCG Database Workshop Summary and Proposal for the First Distributed Production Phase Dirk Duellmann, CERN IT (For the LCG 3D Project :

Slides:



Advertisements
Similar presentations
1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Advertisements

1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
D. Düllmann - IT/DB LCG - POOL Project1 POOL Release Plan for 2003 Dirk Düllmann LCG Application Area Meeting, 5 th March 2003.
CERN - IT Department CH-1211 Genève 23 Switzerland t Relational Databases for the LHC Computing Grid The LCG Distributed Database Deployment.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Database monitoring and service validation Dirk Duellmann CERN IT/PSS and 3D
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
LCG 3D StatusDirk Duellmann1 LCG 3D Throughput Tests Scheduled for May - extended until end of June –Use the production database clusters at tier 1 and.
Update on Database Issues Peter Chochula DCS Workshop, June 21, 2004 Colmar.
Databases Technologies and Distribution Techniques Dirk Duellmann, CERN HEPiX, Rome, April 4th 2006.
LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3.
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
LCG 3D Project Status and Production Plans Dirk Duellmann, CERN IT On behalf of the LCG 3D project CHEP 2006, 15th February, Mumbai.
ATLAS Database Operations Invited talk at the XXI International Symposium on Nuclear Electronics & Computing Varna, Bulgaria, September 2007 Alexandre.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
ATLAS applications and plans LCG Database Deployment and Persistency Workshop 17-Oct-2005,CERN Stefan Stonjek (Oxford), Torre Wenaus (BNL)
Database Administrator RAL Proposed Workshop Goals Dirk Duellmann, CERN.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
3D Workshop Outline & Goals Dirk Düllmann, CERN IT More details at
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
Database Readiness Workshop Summary Dirk Duellmann, CERN IT For the LCG 3D project SC4 / pilot WLCG Service Workshop.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
SC4 Planning Planning for the Initial LCG Service September 2005.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
D. Duellmann - IT/DB LCG - POOL Project1 The LCG Pool Project and ROOT I/O Dirk Duellmann What is Pool? Component Breakdown Status and Plans.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
3D Testing and Monitoring Lee Lueking LCG 3D Meeting Sept. 15, 2005.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
3D Project Status Dirk Duellmann, CERN IT For the LCG 3D project Meeting with LHCC Referees, March 21st 06.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
CERN IT Department CH-1211 Geneva 23 Switzerland t WLCG Operation Coordination Luca Canali (for IT-DB) Oracle Upgrades.
LCG 3D Project Update (given to LCG MB this Monday) Dirk Duellmann CERN IT/PSS and 3D
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
Status of tests in the LCG 3D database testbed Eva Dafonte Pérez LCG Database Deployment and Persistency Workshop.
Database Project Milestones (+ few status slides) Dirk Duellmann, CERN IT-PSS (
D. Duellmann, IT-DB POOL Status1 POOL Persistency Framework - Status after a first year of development Dirk Düllmann, IT-DB.
Replicazione e QoS nella gestione di database grid-oriented Barbara Martelli INFN - CNAF.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
A quick summary and some ideas for the 2005 work plan Dirk Düllmann, CERN IT More details at
CERN - IT Department CH-1211 Genève 23 Switzerland t Service Level & Responsibilities Dirk Düllmann LCG 3D Database Workshop September,
VO Box discussion ATLAS NIKHEF January, 2006 Miguel Branco -
1 LCG Distributed Deployment of Databases A Project Proposal Dirk Düllmann LCG PEB 20th July 2004.
Database Requirements Updates from LHC Experiments WLCG Grid Deployment Board Meeting CERN, Geneva, Switzerland February 7, 2007 Alexandre Vaniachine (Argonne)
Dario Barberis: ATLAS DB S&C Week – 3 December Oracle/Frontier and CondDB Consolidation Dario Barberis Genoa University/INFN.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 6 April 2005.
INFN Tier1/Tier2 Cloud WorkshopCNAF, 22 November 2006 Conditions Database Services How to implement the local replicas at Tier1 and Tier2 sites Andrea.
Database Readiness Workshop Summary Dirk Duellmann, CERN IT For the LCG 3D project GDB meeting, March 8th 06.
Jean-Philippe Baud, IT-GD, CERN November 2007
Status of Distributed Databases (LCG 3D)
Dirk Duellmann CERN IT/PSS and 3D
Database Replication and Monitoring
IT-DB Physics Services Planning for LHC start-up
LCG 3D Distributed Deployment of Databases
Database Services at CERN Status Update
3D Application Tests Application test proposals
Database Readiness Workshop Intro & Goals
LCG Distributed Deployment of Databases A Project Proposal
Conditions Data access using FroNTier Squid cache Server
Workshop Summary Dirk Duellmann.
ATLAS DC2 & Continuous production
Offline framework for conditions data
Presentation transcript:

LCG Database Workshop Summary and Proposal for the First Distributed Production Phase Dirk Duellmann, CERN IT (For the LCG 3D Project : )

LCG 3D Production PhaseDirk Duellmann2 Why a LCG Database Deployment Project? LCG today provides an infrastructure for distributed access to file based data and file replication Physics applications (and grid services) require a similar services for data stored in relational databases –Several applications and services already use RDBMS –Several sites have already experience in providing RDBMS services Goals for common project as part of LCG –increase the availability and scalability of LCG and experiment components –allow applications to access data in a consistent, location independent way –allow to connect existing db services via data replication mechanisms –simplify a shared deployment and administration of this infrastructure during 24 x 7 operation

LCG 3D Production PhaseDirk Duellmann3 M LCG 3D Service Architecture O O O M T1- db back bone - all data replicated - reliable service T2 - local db cache -subset data -only local service T3/4 MM T0 - autonomous reliable service Oracle Streams Cross vendor copy MySQL/SQLight Files Proxy Cache

LCG 3D Production PhaseDirk Duellmann4 3rd LCG Database Workshop Three days workshop Oct with LHC experiments and –Sites: ASCC, CERN, CNAF, BNL, FNAL, GridKA, IN2P3, RAL Contact with PIC and NIKHEF/SARA –Details: Focus on preparation for large scale deployment –Capture current deployment plans –Discuss proposed services and impact on applications –Continue service validation with real apps/workload –Define and schedule production setup at LCG Tier sites

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek5 ATLAS Database / Data Management Project Responsible for DB/DM activities across ATLAS: – DBs for detector production & installation, survey, detector geometry – Online configuration, bookkeeping, run conditions – Online and offline calibrations & alignments – Event data and metadata – Offline processing configuration and bookkeeping – Distributed data management (file-based data) – Distributed database infrastructure and services All of these rely on DB services at CERN and at external institutes, and a distributed DB infrastructure knitting these together Consequently we strongly support and rely on the 3D project! Subprojects Production DBs Installation DBs Geometry DB Online DBs Calib/Align Conditions DB Event Data Distributed DM Offl Processing DB Services SW Support

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek8 Online/T0 DB System Architecture

LCG 3D Production PhaseDirk Duellmann9 Experiment Deployment Plans LHCb –Databases online / Tier 0 / Tier 1 –Oracle streams for conditions replication (COOL) ATLAS –Databases online / Tier 0 / Tier 1 / T2 (mysql) –Oracle streams for conditions replication (COOL) Both interested in FroNtier cache for conditions data in COOL

LCG 3D Production Phase17 October 2005 Latchezar BetevDatabases in ALICE 10 Offline conditions DB relations DAQ Trigger DCS ECS Physics data DCDB AliEn/GLite: metadata file store calibration procedures calibration files AliRoot Calibration classes API files API HLT

LCG 3D Production PhaseLee Lueking11 CMS Online  Offline DB POOL-ORA format Sub-Detector designed format

LCG 3D Production PhaseLee Lueking12 Rough Estimates of CMS DB Resources March through September 2006 Area (CMS contact) Disk Maximum (GB) Concurrent Users Peak usage Transactions Peak usage (Hz) Online P5 (F. Glege) Offline Conditions Tier-0 (L. Lueking/O.Buchmuller) 500 (DB) 100 (per Squid) 2-3 Squids 10 (DB)* 10 (Squid) * Incl. On2off Xfer 10 (DB)* 10 (Squid) * Incl. On2off Xfer Offline Conditions Tier-1 (each site) (L.Lueking/O. Buchmuller) 100 (per Squid) 2-3 Squids/site 10(per Squid) >100  all sites 10 per (Squid) >100  all sites Offline DMS Tier-0 (P. Elmer/L. Turra) 2010

LCG 3D Production PhaseLee Lueking13 Offline FroNTier Resources/Deployment Tier-0: 2-3 Redundant FroNTier servers. Tier-1: 2-3 Redundant Squid servers. Tier-N: 1-2 Squid Servers. Typical Squid server requirements: –CPU/MEM/DISK/NIC=1GHz/1 GB/100GB/Gbit –Network: visible to Worker LAN (private network) and WAN (internet) –Firewall: Two Ports open for URI (FroNTier Launchpad) access and SNMP monitoring (typically 8000 and 3401 respectively) Squid non-requirements –Special hardware (although high-throughput Disk I/O is good) –Cache backup (if disk dies or is corrupted, start from scratch and reload automatically) Squid is easy to install and requires little on-going administration. Squid(s) Tomcat(s) Squid DB Squid Tier 0 Tier 1 Tier N FroNTier Launchpad http JDBC

LCG 3D Production PhaseDirk Duellmann14 Experiment Deployment Plans ALICE –Databases online / Tier 0 - files Tier 1 and higher –Alice s/w for online to offline copy/transformation –Files for conditions data distribution –No 3D service request outside T0 CMS –Databases online / Tier 0 - DB cache Tier 1 and higher –Oracle streams for online to offline replication –FroNtier/POOL for conditions data distribution (cache at T1/T2) Oracle streams as fallback

LCG 3D Production PhaseDirk Duellmann15 Databases in SC3 Middleware Deployment Takes place already for services used in SC3 –Existing setups at the sites –Existing experience with SC workloads -> extrapolate to real production LFC, FTS - Tier 0 and above –Low volume but high availability requirements –CERN: Run on 2-node Oracle cluster; outside Oracle or MySQL CASTOR 2 - CERN and some T1 sites –Scalability requirements may need effort on DB side –CERN: Currently run on 3 Oracle servers Currently not driving the requirements for the database service Consolidation of databases h/w and procedures may reduce effort/diversity at Tier 1 sites

LCG 3D Production PhaseDirk Duellmann16 DB Requirements Status Experiment data at Tier 1 and higher read-only –Few well defined accounts write online/Tier 0 Volume, CPU and transaction estimates still evolving –Need detector s/w integration for conditions data –Need production service in place for SC4 based on today’s estimates –Review access pattern after eg first 3 month of production Iterate via application validation tests –Significant testing effort and flexibility from all participants Experiments, db service and s/w development teams

LCG 3D Production PhaseMaria Girone17 Service CERN The Physics Database services at CERN moved to database clusters

LCG 3D Production PhaseMaria Girone18 CERN Database h/w Evolution Current State ALIC E ATLASCMSLHCbGrid3DNon-LHCValidation - 2-node offline 2-node -- 2x2-node 2-node online test Proposed structure in node4-nodes 2-node2-node (PDB replacement) 2-node valid/test 2-node pilot Non-LHC Ramp up of hardware resources during

LCG 3D Production PhaseDirk Duellmann19 Database Clusters Several sites are testing/deploying Oracle clusters –CERN, CNAF, BNL, FNAL, GridKA, IN2P3, RAL Several experiments foresee Oracle clusters for online systems Focus on db clusters as main building block also for Tier 1 3D will organize detailed DBA level discussions on database cluster setup –Tier 1 DB teams and Online DB teams –Share test plans and expertise among LCG sites and experiments –Cluster setup and existing test results –Storage configuration and performance tests

LCG 3D Production PhaseDirk Duellmann20 Proposed Tier 1 Service Setup Propose to setup for first 6 month –2/3 dual-cpu database nodes with 2GB or more Setup as RAC cluster (preferably) per experiment ATLAS: 3 nodes with 300GB storage (after mirroring) LHCb: 2 nodes with 100GB storage (after mirroring) Shared storage (eg FibreChannel) proposed to allow for clustering –2-3 dual-cpu Squid nodes with 1GB or more Squid s/w packaged by CMS will be provided by 3D 100GB storage per node Need to clarify service responsibility (DB or admin team?) Target s/w release: Oracle 10gR2 –RedHat Enterprise Server to insure Oracle support

LCG 3D Production PhaseDirk Duellmann21 Proposed 2006 Setup Schedule November: h/w setup defined and plan to PEB/GDB January: h/w acceptance tests, RAC setup Begin February: Tier 1 DB readiness workshop February: Apps and streams setup at Tier 0 March: Tier 1 service starts End May: Service review -> h/w defined for full production September: Full LCG database service in place

LCG 3D Production PhaseDirk Duellmann22 Summary Database applications s/w and distribution models firming up - driving application “Conditions database” –ATLAS, CMS and LHCb require access to database data also outside CERN, ALICE only at CERN Two target distribution technologies (Streams and FroNtier) and complementary deployment plans for initial production phase –Online -> offline replication based on streams for all three experiments Propose to setup pre-production services for March 06 and full LHC setup after 6 month deployment experience –Definition of concrete conditions data models in experiment s/w should be aligned

Questions?

Additional Slides

LCG 3D Production PhaseDirk Duellmann25 Tier 2 Setup Only little effort from 3D so far –Will change with full slice test for COOL now BNL / ATLAS have most deployment experience Propose to define and document standard s/w installation –Need more participation of prototype Tier 2 sites

LCG 3D Production PhaseDirk Duellmann26 Project Non-Goals  Store all database data Experiments are free to deploy databases and distribute data under their responsibility  Setup a single monolithic distributed database system Given constraints like WAN connections one can not assume that a single synchronously updated database work and provide sufficient availability.  Setup a single vendor system Technology independence and a multi-vendor implementation will be required to minimize the long term risks and to adapt to the different requirements/constraints on different tiers.  Impose a CERN centric infrastructure to participating sites CERN is one equal partner of other LCG sites on each tier  Decide on an architecture, implementation, new services, policies Produce a technical proposal for all of those to LCG PEB/GDB

LCG 3D Production PhaseDirk Duellmann27 LCG Certificate Support Easier for 3-tier apps Still an issue for client-server db applications –X509 authentication in Oracle and MySQL –Proxy certs can be made work with MySQL but fail so far with Oracle –Authorization (mapping of VOMS to DB roles) still missing Stop-gap solution (read-only access outside T0) acceptable for security and experiments for ~6month –Monitoring will important and needs real user id (DN) Little manpower left in the project –ASCC contributing in the area of Oracle security infrastructure setup –Need a test system and review CERN

LCG 3D Production PhaseDirk Duellmann28 Physics DB CERN Database clusters (RAC) are in production now !! –Big effort of Maria’s team in parallel to ongoing services and significant support load Hope to re-focus resources to consultancy and service improvement –Standardized account setup for smother transition between service levels and more stable client side connection information –Proposed new storage organization provides improved performance and uses available disk volume to decrease recovery latency –Backup policy review and read-only data are important to insure that tape backup volume stays scalable as volume grows –DB Server monitoring for developers/deployment integrated into LEMON –Unavailability caused by security patches and overload are most significant - applications and services need to retry/failover

LCG 3D Production PhaseMaria Girone29 Database Service Levels Development Service (run by IT-DES) –Code development, no large data volumes, limited number of concurrent connections –Once stable, the application code and schema move to validation Validation Service (for key apps) –Sufficient resources for larger tests and optimisation –Allocated together with DBA resources consultancy Needs to be planned in advance –Limited time slots of about 3 weeks Production Service –Full production quality service, including backup, monitoring services, on call intervention procedures –Monitoring to detect new resource consuming applications or changes in access patterns OS level support provided by IT-FIO

LCG 3D Production PhaseDirk Duellmann30 Streams Tested successfully for simple workloads (file catalogs) –LFC replication test requested by LHCb scheduled Several validation test ongoing for conditions data –Online->offline for ATLAS and CMS (CC->CC, CMS P5->CC) –Offline->T1 for ATLAS and LHCb (CERN->RAL->Oxford) Several tests scheduled requiring service and consultancy –0.5 FTE DBA support the CERN side likely to become a bottleneck

LCG 3D Production PhaseDirk Duellmann31 FroNtier CMS Baseline for conditions data  No database service outside T0 required  Simpler squid service on T1 and T2 Integrated with LCG persistency framework via transparent plug-in –Early testing phase in CMS –Interest from other experiments (ATLAS/LHCb) –FroNtier test setup available in 3D test bed

LCG 3D Production PhaseLee Lueking32 CMS Non-event Data Model Conditions data: (Calibration, Alignment, Geometry, Configuration, Monitoring) –Online DB: Schemas designed specific to sub-system; Oracle DB server at P5 –Offline DB: POOL-ORA repositories HLT Farm: Oracle DB server at P5 Tier-0:Oracle DB server at IT –Online to offline: Transform online format to POOL-ORA payloads. Transfer POOL-ORA payloads to offline via Oracle Streams. –Offline (Tier-0) to Tier-N: Plan A: CMS Client POOL-RAL  SQUID proxy/caching servers  FroNTier “pass through” server  POOL-ORA repository. Plan B: Oracle replication to Tier-1 sites, if Plan A is insufficient or fails. Event Data Management System (DMS) –Tier-0: Dataset Bookkeeping Service (DBS), Dataset Location Service (DLS) –Tier-1: Local DBS and DLS for internal bookkeeping. Possible use of Oracle replication w/ Tier-0 if needed for performance/availability. –Tier-2: Local DBS and DLS for internal bookkeeping (not Oracle).

ATLAS applications and plans LCG Database Deployment and Persistency Workshop 17-Oct-2005,CERN Stefan Stonjek (Oxford), Torre Wenaus (BNL)

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek34 Online Databases ATLAS Online uses standardized DB tools from the DB/DM project (and mostly from LCG AA) RAL used as standard DB interface, either directly or indirectly (COOL) – Direct usages: L1 trigger configuration, TDAQ OKS configuration system COOL conditions DB is successor to Lisbon DB for time dependent conditions, configuration data – Links between COOL and online (PVSS, information system) in place – PVSS data sent to PVSS-Oracle and then to COOL (CERN Oracle) Strategy of DB access via ‘offline’ tools (COOL/POOL) and via direct access to the back end DB popular in online Joint IT/ATLAS project to test online Oracle DB strategy being established – Online Oracle DB physically resident at IT, on ATLAS-online secure subnet – Data exported from there to offline central Oracle DB, also IT resident

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek35 Conditions DB – COOL (1) COOL used for conditions data across ATLAS subdetectors and online – ATLAS is the main experiment contributor to its development – Initiative started to use COOL to manage time- dependent aspects of subsystem configurations COOL now fully integrated with ATLAS offline framework Athena Deployment plan written and well received within ATLAS and by IT Usage model: write centrally (CERN Oracle), read mainly via caches/replicas Scalability and performance testing/debugging of COOL and underlying DB services being tested with realistic ATLAS online workloads

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek36 Some ATLAS DB Concerns (1) Scalable distributed access to conditions data – COOL copy/extraction tools in development, and in good hands; will come but aren’t there yet – FroNTier approach of great interest but untouched in ATLAS for lack of manpower FroNTier/RAL integration is welcome, we need to look at it! – DDM already deployed in a limited way for calibration data file management, but needs to be scaled up and the divide between file- and DB-based conditions data better understood – Role of object relational POOL and implications for distributed access still to be understood

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek37 Conclusion (1) ATLAS DB/DM is well aligned with, draws heavily from, and contributes to LCG 3D and LCG AA/persistency; we depend on them being strongly supported and will continue to support them as best we can The ‘easier’ applications from the DB services point of view are well established in production, reasonably well understood, and relatively light in their service/distribution requirements – TC databases (except survey), geometry database

LCG 3D Production PhaseTorre Wenaus, Stefan Stonjek38 Conclusion (2) For the most critical of the rest, the ‘final’ applications now exist in various states of maturity, but more scale/usage information and operational experience is needed before we can be reliably concrete – Conditions DB, event tags?, production DB, DDM For the remainder, applications and even strategies are still immature to non-existent (largely because they relate to the still-evolving analysis model) – Event processing metadata, event tags?, physics dataset selection, provenance metadata

Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005

LCG 3D Production Phase17 October 2005 Latchezar BetevDatabases in ALICE 45 Databases relations and connectivity  Examples from previous slides:  DCDB – “quasi” distributed system, central repository is read- only, updates are infrequent  DAQ – completely closed central system  The biggest challenges remain the Grid file catalogue and offline conditions databases:  Used in a heterogeneous and often ‘hostile’ computing environment (the Grid).  Contains data from wide variety of sources

LCG 3D Production Phase17 October 2005 Latchezar BetevDatabases in ALICE 46 Considerations for Conditions DB  Sources for conditions DB and relations to other DBs are already quite complicated:  All databases potentially containing conditions information are “closed”, i.e. only accessible at CERN  It would be difficult to provide access methods to all DBs from the production and analysis code  ALICE uses ROOT as offline framework base:  Naturally defines the technology choice for object store – all conditions data are stored as root files  Additional condition - these files are read-only

LCG 3D Production Phase17 October 2005 Latchezar BetevDatabases in ALICE 47 Considerations for Conditions DB (2)  Conditions data should be accessible in a Grid distributed production and analysis environment  The root files are registered in the Grid Distributed File Catalogue:  No need for distributed DBMS in a traditional sense and with all accompanying problems (access, replication, authentication) – a very big plus  Assures worldwide access to all files and associated tags  Drawbacks:  Replication of information  Depends almost entirely on the Grid file catalogue functionality

LCG 3D Production Phase17 October 2005 Latchezar BetevDatabases in ALICE 48 Medium term plans  Grid file catalogue is used routinely in ALICE since several years in the scope of the physics data challenges (PDC04, PDC05, SC3)  ROOT Grid access classes are now mature  AliRoot Conditions DB access framework is complete  Beginning of 2006 – Test of ALICE TPC with operational DCS and DAQ:  Test of both DBs and in addition the offline Conditions DB  Beginning of 2006 – Physics Data Challenge ’06:  Final test of the entire ALICE offline framework on the Grid, including test of Conditions framework

LCG 3D Production Phase17 October 2005 Latchezar BetevDatabases in ALICE 49 Conclusions  ALICE has a rich field of databases used in the various groups for wide variety of tasks  The development of the majority of DBs is well advanced:  Some of them are already in production since several years  The biggest challenge remaining is to gather the information from all sources and make it available for Grid data reconstruction and analysis:  In this context, ALICE has decided to re-use the already existing Grid file catalogue technology  And enrich the AliRoot framework with tools allowing connectivity to all other DBs currently in existence