ATLAS Metrics for CCRC’08 Database Milestones WLCG CCRC'08 Post-Mortem Workshop CERN, Geneva, Switzerland June 12-13, 2008 Alexandre Vaniachine.

Slides:



Advertisements
Similar presentations
Tales from the Lab: Experiences and Methodology Demand Technology User Group December 5, 2005 Ellen Friedman SRM Associates, Ltd.
Advertisements

ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
DB server limits (process/sessions) Carlos Fernando Gamboa, BNL Andrew Wong, TRIUMF WLCG Collaboration Workshop, CERN Geneva, April 2008.
CERN - IT Department CH-1211 Genève 23 Switzerland t Relational Databases for the LHC Computing Grid The LCG Distributed Database Deployment.
ATLAS computing in Geneva Szymon Gadomski, NDGF meeting, September 2009 S. Gadomski, ”ATLAS computing in Geneva", NDGF, Sept 091 the Geneva ATLAS Tier-3.
Load Test Planning Especially with HP LoadRunner >>>>>>>>>>>>>>>>>>>>>>
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
LCG 3D StatusDirk Duellmann1 LCG 3D Throughput Tests Scheduled for May - extended until end of June –Use the production database clusters at tier 1 and.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
1 An SLA-Oriented Capacity Planning Tool for Streaming Media Services Lucy Cherkasova, Wenting Tang, and Sharad Singhal HPLabs,USA.
SC4 Workshop Outline (Strong overlap with POW!) 1.Get data rates at all Tier1s up to MoU Values Recent re-run shows the way! (More on next slides…) 2.Re-deploy.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Alexei Klimentov : ATLAS Computing CHEP March Prague Reprocessing LHC beam and cosmic ray data with the ATLAS distributed Production System.
CCRC08-1 report WLCG Workshop, April KorsBos, ATLAS/NIKHEF/CERN.
ATLAS Scalability Tests of Tier-1 Database Replicas WLCG Collaboration Workshop (Tier0/Tier1/Tier2) Victoria, British Columbia, Canada September 1-2, 2007.
Workshop Summary (my impressions at least) Dirk Duellmann, CERN IT LCG Database Deployment & Persistency Workshop.
ATLAS Database Operations Invited talk at the XXI International Symposium on Nuclear Electronics & Computing Varna, Bulgaria, September 2007 Alexandre.
OSG Area Coordinator’s Report: Workload Management April 20 th, 2011 Maxim Potekhin BNL
2nd September Richard Hawkings / Paul Laycock Conditions data handling in FDR2c  Tag hierarchies set up (largely by Paul) and communicated in advance.
Databases E. Leonardi, P. Valente. Conditions DB Conditions=Dynamic parameters non-event time-varying Conditions database (CondDB) General definition:
John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
Database Administrator RAL Proposed Workshop Goals Dirk Duellmann, CERN.
CERN Physics Database Services and Plans Maria Girone, CERN-IT
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Online-Offsite Connectivity Experiments Catalin Meirosu *, Richard Hughes-Jones ** * CERN and Politehnica University of Bucuresti ** University of Manchester.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle Real Application Clusters (RAC) Techniques for implementing & running robust.
CERN - IT Department CH-1211 Genève 23 Switzerland t COOL Conditions Database for the LHC Experiments Development and Deployment Status Andrea.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
LCG CCRC’08 Status WLCG Management Board November 27 th 2007
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
CCRC’08 Monthly Update ~~~ WLCG Grid Deployment Board, 14 th May 2008 Are we having fun yet?
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Development, Deployment and Operations of ATLAS Databases XVI International Conference on Computing in High Energy and Nuclear Physics Victoria, British.
CERN IT Department CH-1211 Genève 23 Switzerland t Streams Service Review Distributed Database Workshop CERN, 27 th November 2009 Eva Dafonte.
CERN IT Department CH-1211 Geneva 23 Switzerland t WLCG Operation Coordination Luca Canali (for IT-DB) Oracle Upgrades.
WLCG Service Report ~~~ WLCG Management Board, 16 th September 2008 Minutes from daily meetings.
LCG 3D Project Update (given to LCG MB this Monday) Dirk Duellmann CERN IT/PSS and 3D
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
Grid Technologies for Distributed Database Services 3D Project Meeting CERN, May 19, 2005 A. Vaniachine (ANL)
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
CERN IT Department CH-1211 Geneva 23 Switzerland t Distributed Database Operations Workshop CERN, 17th November 2010 Dawid Wójcik Streams.
Database Project Milestones (+ few status slides) Dirk Duellmann, CERN IT-PSS (
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
HEPiX Rome – April 2006 The High Energy Data Pump A Survey of State-of-the-Art Hardware & Software Solutions Martin Gasthuber / DESY Graeme Stewart / Glasgow.
Summary of SC4 Disk-Disk Transfers LCG MB, April Jamie Shiers, CERN.
A quick summary and some ideas for the 2005 work plan Dirk Düllmann, CERN IT More details at
DB Questions and Answers open session (comments during session) WLCG Collaboration Workshop, CERN Geneva, 24 of April 2008.
Database Requirements Updates from LHC Experiments WLCG Grid Deployment Board Meeting CERN, Geneva, Switzerland February 7, 2007 Alexandre Vaniachine (Argonne)
Dario Barberis: ATLAS DB S&C Week – 3 December Oracle/Frontier and CondDB Consolidation Dario Barberis Genoa University/INFN.
J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April ATLAS Grid Activities Preparing for Data Analysis Jim Shank.
INFN Tier1/Tier2 Cloud WorkshopCNAF, 22 November 2006 Conditions Database Services How to implement the local replicas at Tier1 and Tier2 sites Andrea.
Computing Operations Roadmap
Virtualization and Clouds ATLAS position
IT-DB Physics Services Planning for LHC start-up
3D Application Tests Application test proposals
Elizabeth Gallas - Oxford ADC Weekly September 13, 2011
Database Readiness Workshop Intro & Goals
WLCG Service Interventions
Conditions DB Data Distribution
Workshop Summary Dirk Duellmann.
3D Project Status Report
Presentation transcript:

ATLAS Metrics for CCRC’08 Database Milestones WLCG CCRC'08 Post-Mortem Workshop CERN, Geneva, Switzerland June 12-13, 2008 Alexandre Vaniachine

2 WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, D ATLAS Milestone Accomplishment 3D database services at Tier-1 sties are critical for data reprocessing –Since data reprocessing depends not only on database access it was launched at the end of the CCRC’08 exercise 3D servers at the Tier-1 sites were asked to operate at the milestone metrics, which is the sessions load from ATLAS data reprocessing jobs In a steady-state LHC data processing operation we require capacities capable of sustaining ~1k concurrent Oracle sessions in total –During CCRC’08 3D services demonstrated capability to exceed that cumulative ATLAS Milestone target by 1/3 That milestone was achieved, with some of Tier-1 sites not yet tested: –Six sites operated above the ATLAS milestone metrics two sites operated close to ¾ of the metrics – two remaining sites were tested at less than ½ of the metrics Based on the CCRC’08 experience to date, we are confident that critical WLCG 3D services are ready for ATLAS LHC data processing

3 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 What Else Has Been Accomplished During CCRC’08 3D streamed to Tier-1 sites all the COOL data needed for bulk reprocessing –All ATLAS data reprocessing jobs got the data replicated to Tier-1s Recently upgraded RAC hardware at CERN delivered enough capacities to handle unexpected workload from FDR-2 data processing at Tier-0 At the tested Tier-1 sites reading of the streamed DCS data using latest COOL 2.4 software release worked with good performance –The DCS data need special attention because of the largest volumes “Upstream” 3D streams replication has been setup from Muon Calibration Centers in Rome, Michigan and Munich to INTR server at CERN –Used during ATLAS FDR-2 exercise during CCRC’08 Critical for ATLAS “24-h calibration loop” Database monitoring worked well

4 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 Monitoring Burst Load at Tier-1s ATLAS does not have spare WLCG 3D Oracle capacities deployed to support burst loads when many data reprocessing jobs start at once on an empty cluster Short burst periods correspond to submission of limited number of test jobs for data reprocessing tasks TRIUMF CPU load at peak: 98% peak for 1 min 70% average Note: Summary load on both Oracle RAC nodes at TRIUMF Tier-1

5 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 Monitoring CCRC’08 Bulk Data Reprocessing at Tier-1 Typical load from bulk data re-processing (5k jobs) Max load happens for burst periods when many jobs start simultaneously Note: Summary load on both Oracle RAC nodes at FZK Tier-1

6 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 Monitoring CCRC’08 Cosmics Data Reprocessing at Tier-1* Load from US ATLAS users work * Reprocessing using large ATLAS Tier-2 sites in the U.S. in addition Tier-1 resources at BNL Note: Sessions load from both Oracle RAC nodes at BNL Tier-1, CPU load from a single node Load from bulk data re- processing

7 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 CCRC’08 Monitoring of FDR-2 Data Processing at Tier-0 Start of bulk data processing Express stream data processing Load from ATLAS users work One-run data processing (discontinued) Daily data-quality assessment meeting to "sign- off" DB & S/W tags for bulk data processing

8 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 DCS Scalability Test Load at CNAF Tier-1 Note: Summary load on three Oracle RAC nodes at CNAF Tier-1 ATLAS Conditions DB scalability tests at CNAF with new COOL 2.4 software show good results for large DCS data reads 30% CPU load at max sessions shows that ATLAS capacities request was not excessive

9 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 CCRC’08 Post-Mortem Findings Latest scalability test results at CNAF show that WLCG 3D capacities deployed on ATLAS request are not excessive –For example, ATLAS does not have spare WLCG 3D Oracle capacities deployed to support burst loads when many data reprocessing jobs start at once on an empty cluster Oracle connect strings changes should be planned and announced well in advance, so that they can be propagated to ATLAS jobs configurations Evaluation of bulk data reprocessing at the Tier-2 sites uncovered new requirements: –Deployment of additional Oracle hardware capacities at BNL Tier-1 may be required to support bulk reprocessing of LHC data at large Tier-2 sites in the U.S. –How to connect to remote Oracle server via the site firewall?

10 Alexandre Vaniachine WLCG CCRC'08 Post-Mortem Workshop, CERN, June 12-13, 2008 Conclusions and Next Steps The WLCG 3D services deployed for ATLAS demonstrated capability to exceed cumulative ATLAS Milestone target by 1/3 –Based on the CCRC’08 experience to date, we are confident that critical WLCG 3D services are ready for ATLAS LHC data processing Since there is little room for an error in our estimates of 3D capacities required for ATLAS –We have to validate these estimates using latest ATLAS software and Computing Model ATLAS validation of 3D services will continue with further data reprocessing exercises increasing in scope and complexity: –We plan to complete testing using latest FDR-2 data samples –We will test all Tier-1 sites using access to the DCS data Expect higher load from new DCS tests at all Tier-1