LCG 3D Project Status For more details: Dirk Düllmann, CERN IT-ADC LCG Application Area Meeting 6 April, 2005.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel - WP2. WP2 Overview.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Database Architectures and the Web
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Andrew McNab - EDG Access Control - 14 Jan 2003 EU DataGrid security with GSI and Globus Andrew McNab University of Manchester
The POOL Persistency Framework POOL Summary and Plans.
Objektorienteret Middleware Presentation 2: Distributed Systems – A brush up, and relations to Middleware, Heterogeneity & Transparency.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
D. Düllmann - IT/DB LCG - POOL Project1 POOL Release Plan for 2003 Dirk Düllmann LCG Application Area Meeting, 5 th March 2003.
Database monitoring and service validation Dirk Duellmann CERN IT/PSS and 3D
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Microsoft Active Directory(AD) A presentation by Robert, Jasmine, Val and Scott IMT546 December 11, 2004.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
RLS Tier-1 Deployment James Casey, PPARC-LCG Fellow, CERN 10 th GridPP Meeting, CERN, 3 rd June 2004.
Databases Technologies and Distribution Techniques Dirk Duellmann, CERN HEPiX, Rome, April 4th 2006.
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
Responsibilities of ROC and CIC in EGEE infrastructure A.Kryukov, SINP MSU, CIC Manager Yu.Lazin, IHEP, ROC Manager
ATLAS Detector Description Database Vakho Tsulaia University of Pittsburgh 3D workshop, CERN 14-Dec-2004.
Database Administrator RAL Proposed Workshop Goals Dirk Duellmann, CERN.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
3D Workshop Outline & Goals Dirk Düllmann, CERN IT More details at
Feedback from the POOL Project User Feedback from the POOL Project Dirk Düllmann, LCG-POOL LCG Application Area Internal Review October 2003.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
G.Govi CERN/IT-DB 1 September 26, 2003 POOL Integration, Testing and Release Procedure Integration  Packages structure  External dependencies  Configuration.
Database authentication in CORAL and COOL Database authentication in CORAL and COOL Giacomo Govi Giacomo Govi CERN IT/PSS CERN IT/PSS On behalf of the.
D. Duellmann - IT/DB LCG - POOL Project1 The LCG Pool Project and ROOT I/O Dirk Duellmann What is Pool? Component Breakdown Status and Plans.
Some Ideas for a Revised Requirement List Dirk Duellmann.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
LCG Distributed Databases Deployment – Kickoff Workshop Dec Database Lookup Service Kuba Zajączkowski Chi-Wei Wang.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
LCG 3D Project Update (given to LCG MB this Monday) Dirk Duellmann CERN IT/PSS and 3D
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
Status of tests in the LCG 3D database testbed Eva Dafonte Pérez LCG Database Deployment and Persistency Workshop.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Database Project Milestones (+ few status slides) Dirk Duellmann, CERN IT-PSS (
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
D. Duellmann, IT-DB POOL Status1 POOL Persistency Framework - Status after a first year of development Dirk Düllmann, IT-DB.
Replicazione e QoS nella gestione di database grid-oriented Barbara Martelli INFN - CNAF.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
A quick summary and some ideas for the 2005 work plan Dirk Düllmann, CERN IT More details at
CERN - IT Department CH-1211 Genève 23 Switzerland t Service Level & Responsibilities Dirk Düllmann LCG 3D Database Workshop September,
DB Questions and Answers open session (comments during session) WLCG Collaboration Workshop, CERN Geneva, 24 of April 2008.
Grid Deployment Technical Working Groups: Middleware selection AAA,security Resource scheduling Operations User Support GDB Grid Deployment Resource planning,
1 LCG Distributed Deployment of Databases A Project Proposal Dirk Düllmann LCG PEB 20th July 2004.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 6 April 2005.
Jean-Philippe Baud, IT-GD, CERN November 2007
Dirk Duellmann CERN IT/PSS and 3D
(on behalf of the POOL team)
LCG 3D Distributed Deployment of Databases
3D Application Tests Application test proposals
Database Readiness Workshop Intro & Goals
LCG Distributed Deployment of Databases A Project Proposal
Dirk Düllmann CERN Openlab storage workshop 17th March 2003
Workshop Summary Dirk Duellmann.
POOL/RLS Experience Current CMS Data Challenges shows clear problems wrt to the use of RLS Partially due to the normal “learning curve” on all sides in.
Leigh Grundhoefer Indiana University
Presentation transcript:

LCG 3D Project Status For more details: Dirk Düllmann, CERN IT-ADC LCG Application Area Meeting 6 April, 2005

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN2 Distributed Deployment of Databases LCG provides infrastructure for storage, distribution and replication of file based dataLCG provides infrastructure for storage, distribution and replication of file based data Physics applications (and grid m/w) require a similar services for data hosted in relational databasesPhysics applications (and grid m/w) require a similar services for data hosted in relational databases –Several applications and grid services use RDBMS - and more are coming this year –Several sites have already experience in providing RDBMS services Goals for the 3D project as part of LCGGoals for the 3D project as part of LCG –provide LCG apps with a consistent, location independent access to database services –increase the availability and scalability of database applications via replication –arrange for shared deployment and administration of this infrastructure during 24 x 7 operation Joint project between LCG sites, experiments and s/w projectsJoint project between LCG sites, experiments and s/w projects –Time frame for first service: deployment in autumn this year –Evolve together with computing models and requirements towards LHC startup Held Distributed Database workshop at CERN last DecemberHeld Distributed Database workshop at CERN last December

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN3 3D Non-Goals  Store all database data Experiments are free to deploy databases and distribute data under their responsibility.  Service only for distribution data Need a database service also for data which is site local or distributed via application specific methods Even though many specific methods will make deployment difficult  Setup a single monolithic distributed database system Given constraints like WAN connections one can not assume that a single synchronously updated database could work and provide sufficient availability.  Setup a single vendor system Technology independence and a multi-vendor implementation will be required to minimize the long term risks and to adapt to the different requirements/constraints on different tiers.  Impose a CERN centric infrastructure on participating sites CERN is one equal partner of other LCG sites on each tier  Decide on an architecture, implementation, new services, policies Produce a technical proposal for all of those to LCG PEB/GDB

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN4 Project Structure WP1 -Data Inventory and Distribution Requirements –Members are s/w developers from experiments and grid services that use RDBMS data –Gather data properties (volume, ownership) and requirements –Integrate access to 3D services into their software WP2 - Database Service Definition and Implementation –Members are technology and deployment experts from LCG sites –Propose a deployment implementation and common deployment procedures WP3 - Evaluation Tasks –Short, well defined technology evaluations against the requirements delivered by WP1 –Evaluation are proposed by WP2 (evaluation plan) and typically executed by the people proposing a technology for the service implementation and result in a short evaluation report

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN5 3D Data Inventory Collect and maintain a catalog of main RDBMS data types and their service requirementsCollect and maintain a catalog of main RDBMS data types and their service requirements Experiments and grid s/w projects provide a table for each data type which is candidate for storage and replication via the 3D serviceExperiments and grid s/w projects provide a table for each data type which is candidate for storage and replication via the 3D service –Basic storage properties Data description, expected volume on T0/1/2 in 2005 (and evolution) Ownership model: read-only, single user update, single site update, concurrent update –Replication/Caching properties Replication model: site local, all t1, sliced t1, all t2, sliced t2 … Consistency/Latency: how quickly do changes need to reach other sites/tiers Application constraints: DB vendor and DB version constraints –Reliability and Availability requirements Essential for whole grid operation, for site operation, for experiment production, Backup and Recovery policy –acceptable time to recover, location of backup(s)

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN6 Requirement Gathering All participating experiments (ATLAS, CMS and LHCb) have signed up representativesAll participating experiments (ATLAS, CMS and LHCb) have signed up representatives –Rather 3 than 2 names –Tough job, as survey inside experiment crosses many boundaries! Online vs offline, apps development vs production teams Started with a simple spreadsheet capturing the various applicationsStarted with a simple spreadsheet capturing the various applications –Grouped by application One line per replication and Tier (distribution step) –Two main patterns Experiment data: data fan-out - T0->Tn Grid services: data consolidation - Tn->T0 –But some exceptions which need to be documented… Aim to collect complete set of requirements for database servicesAim to collect complete set of requirements for database services –Eg also online data or data which is stored locally but never leaves a tier –Needed to properly size the h/w at each site

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN8 Preliminary Summary for 2005 Applications mentioned so farApplications mentioned so far –FileCatalog, Conditions, Geomentry, Bookkeeping, Physics Meta Data, Collections, Grid Monitoring, TransferDB –Suspect that several smaller(?) applications are still missing Total volume per experiment: GBTotal volume per experiment: GB –Error bars likely as big as the current spread –Significant flexibility/headroom required for service side! Number of applications to be supported: O(5)Number of applications to be supported: O(5) –Many applications under development, many still bound to one db vendor (Oracle or MySQL) –Most distributed applications use either RAL or ODBC/JDBC based implementation Distributed Data becomes read-only down from T0Distributed Data becomes read-only down from T0 –Conservative approach for first service deployment –Current model does not require multi-master replication Please make sure your experiment representatives know about your application requirements!Please make sure your experiment representatives know about your application requirements!

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN9 3D Site Contacts Established contact to several Tier 1 and Tier 2 sitesEstablished contact to several Tier 1 and Tier 2 sites –Tier 1: ASCC, BNL, CERN, CNAF, FNAL, GridKa, IN2P3, RAL –Tier 2: ANL, U Chicago Regular 3D meetingsRegular 3D meetings –Bi-weekly phone/VRVS meetings for Requirement and Service definition WG Visited RAL, FNAL, BNL, IN2P3 and CNAFVisited RAL, FNAL, BNL, IN2P3 and CNAF –Very useful discussions with experiment developers and database service teams there All above sites have expressed interest in project participationAll above sites have expressed interest in project participation –Most sites have allocated and installed h/w and database servers for participation in the 3D test bed BNL has started a Oracle setup for 3D participation –U Chicago agreed to act as Tier 2 in the testbed Contacting DB teams at PIC, NIKHEFContacting DB teams at PIC, NIKHEF

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN10 Database Services at LCG Sites Today Several tier 1 sites provide production level services for HEP and non-HEP applicationsSeveral tier 1 sites provide production level services for HEP and non-HEP applications –Significant deployment experience and well established service exists… –… but can not be changed easily without affecting other site activities Tier 2 sites can only provide very limited manpower for database serviceTier 2 sites can only provide very limited manpower for database service –Part time administration by the same people responsible for fabric –Only a simple, constrained database service should be assumed Database service definition shouldDatabase service definition should –be build on existing experience (eg from –give a clear message about achievable service level at tier2

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN11 Database Service Policies Several sites have deployment policies in placeSeveral sites have deployment policies in place –E.g. FNAL: Staged service levels –Development -> Integration -> Production systems Well defined move of new code / DB schema during development process –Apps developers and DB experts review and optimize schema before production deployment Similar policy proposal prepared for CERN physics database servicesSimilar policy proposal prepared for CERN physics database services –To avoid recent interference between key production applications of different experiments on shared resources Caused by missing indices, inefficient queries, inadequate hardware resources –Storage volume alone is not a sufficient metric to define a database service Need a complete list of applications and a reference workload for each key application to define and optimize the serviceNeed a complete list of applications and a reference workload for each key application to define and optimize the service –How many requests (queries) from how many clients on how much data are required to meet the performance goals? Especially for distributed DB service this will be essential to avoid surprises on either sideEspecially for distributed DB service this will be essential to avoid surprises on either side

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN12 Tier 1 Service Split Discussion only just startingDiscussion only just starting –Draft proposal based on input received from FNAL (Anil Kumar) Local ServicesLocal Services –Server installation, bug & security fixes –OS patches/upgrades –Backup/recovery support –Data migration (between db servers) Shared ServicesShared Services –Db and OS accounts & privileges –Storage support (adding more space) –Monitoring DB alerts, “killer queries cron job output Host system load, space thresholds –Performance problems & optimization Site CommunicationSite Communication –Proposal to setup a shared (web based) Log-Book, mailing lists –Need to establish regular DBA meeting –eg as part of weekly/bi-weekly 3D meetings

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN13 Local Database vs. Local Cache FNAL developed FroNtier systemFNAL developed FroNtier system –a combination of http based database access with proxy caches on the way to the client –Performance gains reduced real database access for largely read-only data reduced transfer overhead compared to low level SOAP RPC based approaches –Deployment gains Web caches (eg squid) are much simpler to deploy than databases and could remove the need for a local database deployment on some tiers No vendor specific database libraries on the client side “Firewall friendly” tunneling of requests through a single port Expect cache technology to play a significant roleExpect cache technology to play a significant role –towards the higher tiers which may not have the resources to run a reliable database service

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN14 M Proposed 3D Service Architecture O O O M T1- db back bone - all data replicated - reliable service T2 - local db cache -subset data -only local service T3/4 MM T0 - autonomous reliable service Oracle Streams Cross vendor extract MySQL Files Proxy Cache

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN15 LCG 3D Testbed ASCC, CERN, CNAF, FNAL are connected nowASCC, CERN, CNAF, FNAL are connected now –RAL and BNL expected to join soon –Not the production setup yet, but sufficient for first realistic test loads – Currently testing distributed file catalog workload Two technologies can be tested and comparedTwo technologies can be tested and compared –Oracle 10g test-bed database service CERN has provided Install kits and documentation are provided for test bed sites (but can not offer offsite support) –FroNtier service FroNtier + database server is being setup at CERN Preparing for squid installations at other sites FNAL has prepared s/w packages and took over installation in the testbed Oracle Enterprise Manager installation for test bed administration and server diagnosticsOracle Enterprise Manager installation for test bed administration and server diagnostics Will need to work on database for client side monitoringWill need to work on database for client side monitoring Based on experience gained at RUN2 database deployment

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN16 Possible 3D use in SC3 Timescale for first 3D deployment only after SC3 startTimescale for first 3D deployment only after SC3 start –But will try to arrange for a preproduction service with the participating sites –Need a list of DB applications to be deployed in SC3 Eg a subset of the 3D spreadsheet with reduced volume (and lower number of sites) –Need to coordinate the software integration with the experiment development teams to be ready for SC3 –Need to allocate database resources for SC3 Once the list of candidate apps has been agreed with the experiments:Once the list of candidate apps has been agreed with the experiments: –Either reduced database service at SC3 start –Or join SC3 later as 3D service / 3D database apps become available

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN17 Service Integration with Application Software Distributed database service to be coupled to the user applicationDistributed database service to be coupled to the user application –Use POOL RAL as 3D reference implementation –Use POOL development resources to implement application coupling The 3D project has no real development resources DB service catalog - how to find a suitable database replica?DB service catalog - how to find a suitable database replica? –In contrast to file replicas the DB vendor may depend on the site (tier) on which a job is executed –Avoid hard-coding of physical connection strings..or spreading them as part of configuration files Allow a local service to be transparently relocated to another machine Prototype based on POOL file catalogPrototype based on POOL file catalog –supports network disconnected lookup based on XML files –Final implementation may still change Eg if a suitable grid-service has been identified

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN18 More Application Integration.. Connection Pooling and Error HandlingConnection Pooling and Error Handling –ATLAS developments being integrated into RAL –Consistent handling of error conditions / retry –Connection pooling Client DiagnosticsClient Diagnostics –Collect timing of top queries of the current application in POOL/RAL –Evaluate packages to aggregate diagnostics across many clients Authentication and AuthorizationAuthentication and Authorization –Mapping between grid certificate and database rights Need to support roles in experiments computing model –Plan to evaluate/integrate with package developed by G.Ganis for ROOT and xrootd Most of the above are useful for local or centralized database applications tooMost of the above are useful for local or centralized database applications too

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN19 Authorization Model: Schemata & Roles Start from a generic model for applications, schemata, accounts and roles which can be implemented in MySQL & OracleStart from a generic model for applications, schemata, accounts and roles which can be implemented in MySQL & Oracle ApplicationApplication –one or more schemata which implement a complete user component (eg a condition DB) SchemaSchema –set of database objects (eg tables, views, stored procedures) –which have many internal relations and only few well defined external relations –share a common set of roles RoleRole –set of of access rights to one or more schemata –reader, writer { and eg implicit: schema owner } –Read access may be given by default to all VO members to simplify the VO management (eg VOMS)

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN20 Authentication & Users Authentication: identifies a user to a databaseAuthentication: identifies a user to a database –Connections happen for several applications/schemata: need to acquire the correct mix of roles –Identity is used for authorization and for monitoring Several authentication methods need to be supportedSeveral authentication methods need to be supported –user/password pairs (eg online) Roles are assigned to database user account by the dba Shared accounts/pw are evil, too many db accounts are a deployment problem –grid certificate + VO roles (on a worker node) Roles are assigned by the VO management (eg using VOMS) Admin tools and procedures can be common with file access –Both required? Any overlap? Authentication method needs to be transparent to the applicationAuthentication method needs to be transparent to the application

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN21 Application Deployment Issues Authentication and Role definition should be transparent to applicationsAuthentication and Role definition should be transparent to applications –Authentication should be encapsulated in a service –Roles are not explicitly used inside applications Applications need to be re-locatableApplications need to be re-locatable –Insure that applications (schemata?) can move to another server (eg one with more free resources) What is the granule of movement? –This may requires individual db sessions/connections Need to control the number of concurrent connectionsNeed to control the number of concurrent connections –Eg via connection pooling on the client side Will incorporate these design guidelines in POOL/RAL developmentWill incorporate these design guidelines in POOL/RAL development

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN22 3D Manpower During the first phase of requirement gatheringDuring the first phase of requirement gathering –IT.35 FTE (Dirk) + 1 FTE (Kuba) +.5 FTE (Eva) –Experiments: database link-people –Sites: 1-3 dba contributing to the testbed setup and service definition discussions Discussing with ASCC dba/developers setup of a certificate based db serviceDiscussing with ASCC dba/developers setup of a certificate based db service Will need POOL manpower to implement s/w connectionWill need POOL manpower to implement s/w connection Will need help from experiments to run their application tests in the 3D test bedWill need help from experiments to run their application tests in the 3D test bed –After successful validation at tier 0 Need to make sure that database services get proper priority wrt to other deployment areasNeed to make sure that database services get proper priority wrt to other deployment areas –Experiments should give their input in the relevant places

3D Status and Plans, AA meeting 6. April 2005D.Duellmann, CERN23Summary A distributed database infrastructure promises to provide scalability and availability for database applications at LHCA distributed database infrastructure promises to provide scalability and availability for database applications at LHC –The LCG 3D project has been started as joint project between experiments and LCG sites to coordinate the definition of this service –Several distribution options are available for planning/design of new applications DB Service DefinitionDB Service Definition –Very relevant deployment/development experience from FNAL –Service task split and application validation policy proposals are firming up –Oracle 10g based replication test-bed expands to first T1 Several new DB applications will be developed this yearSeveral new DB applications will be developed this year –Expect high support / consultancy load on database services (especially at Tier 0) for 2005 –Early planning of concrete deployment and distribution models is key for successful production SC3 support from 3D is only now being discussedSC3 support from 3D is only now being discussed –Participation in SC3 would be a very useful first test of 3D services –Concrete time plan needs negotiation with experiments/sites