Efi.uchicago.edu ci.uchicago.edu Data Federation Strategies for ATLAS using XRootD Ilija Vukotic On behalf of the ATLAS Collaboration Computation and Enrico.

Slides:



Advertisements
Similar presentations
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Advertisements

Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Efi.uchicago.edu ci.uchicago.edu FAX update Rob Gardner Computation and Enrico Fermi Institutes University of Chicago Sep 9, 2013.
Efi.uchicago.edu ci.uchicago.edu FAX status report Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago US ATLAS Computing Integration.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
ATLAS DQ2 Deletion Service D.A. Oleynik, A.S. Petrosyan, V. Garonne, S. Campana (on behalf of the ATLAS Collaboration)
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Storage Wahid Bhimji DPM Collaboration : Tasks. Xrootd: Status; Using for Tier2 reading from “Tier3”; Server data mining.
FAX UPDATE 1 ST JULY Discussion points: FAX failover summary and issues Mailing issues Panda re-brokering to sites using FAX cost and access Issue.
FAX UPDATE 26 TH AUGUST Running issues FAX failover Moving to new AMQ server Informing on endpoint status Monitoring developments Monitoring validation.
Efi.uchicago.edu ci.uchicago.edu Towards FAX usability Rob Gardner, Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago US ATLAS.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
Efi.uchicago.edu ci.uchicago.edu FAX meeting intro and news Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Federated Xrootd.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
Efi.uchicago.edu ci.uchicago.edu FAX Dress Rehearsal Status Report Ilija Vukotic on behalf of the atlas-adc-federated-xrootd working group Computation.
PanDA Update Kaushik De Univ. of Texas at Arlington XRootD Workshop, UCSD January 27, 2015.
Efi.uchicago.edu ci.uchicago.edu Using FAX to test intra-US links Ilija Vukotic on behalf of the atlas-adc-federated-xrootd working group Computing Integration.
Efi.uchicago.edu ci.uchicago.edu FAX status developments performance future Rob Gardner Yang Wei Andrew Hanushevsky Ilija Vukotic.
Storage cleaner: deletes files on mass storage systems. It depends on the results of deletion, files can be set in states: deleted or to repeat deletion.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Efi.uchicago.edu ci.uchicago.edu Status of the FAX federation Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Tier 1 /
Network awareness and network as a resource (and its integration with WMS) Artem Petrosyan (University of Texas at Arlington) BigPanDA Workshop, CERN,
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
Efi.uchicago.edu ci.uchicago.edu FAX status report Ilija Vukotic on behalf of the atlas-adc-federated-xrootd working group S&C week Jun 2, 2014.
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
XROOTD AND FEDERATED STORAGE MONITORING CURRENT STATUS AND ISSUES A.Petrosyan, D.Oleynik, J.Andreeva Creating federated data stores for the LHC CC-IN2P3,
FAX PERFORMANCE TIM, Tokyo May PERFORMANCE TIM, TOKYO, MAY 2013ILIJA VUKOTIC 2  Metrics  Data Coverage  Number of users.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
FAX UPDATE 12 TH AUGUST Discussion points: Developments FAX failover monitoring and issues SSB Mailing issues Panda re-brokering to FAX Monitoring.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
Global ADC Job Monitoring Laura Sargsyan (YerPhI).
Efi.uchicago.edu ci.uchicago.edu Ramping up FAX and WAN direct access Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Efi.uchicago.edu ci.uchicago.edu Storage federations, caches & WMS Rob Gardner Computation and Enrico Fermi Institutes University of Chicago BigPanDA Workshop.
Network integration with PanDA Artem Petrosyan PanDA UTA,
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
WLCG Transfers Dashboard A unified monitoring tool for heterogeneous data transfers. Alexandre Beche.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
ATLAS Distributed Computing ATLAS session WLCG pre-CHEP Workshop New York May 19-20, 2012 Alexei Klimentov Stephane Jezequel Ikuo Ueda For ATLAS Distributed.
XRootD Monitoring Report A.Beche D.Giordano. Outlines  Talk 1: XRootD Monitoring Dashboard  Context  Dataflow and deployment model  Database: storage.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
Campana (CERN-IT/SDC), McKee (Michigan) 16 October 2013 Deployment of a WLCG network monitoring infrastructure based on the perfSONAR-PS technology.
DPM in FAX (ATLAS Federation) Wahid Bhimji University of Edinburgh As well as others in the UK, IT and Elsewhere.
Efi.uchicago.edu ci.uchicago.edu FAX splinter session Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Tier 1 / Tier 2 /
Efi.uchicago.edu ci.uchicago.edu Federating ATLAS storage using XrootD (FAX) Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
Efi.uchicago.edu ci.uchicago.edu Sharing Network Resources Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago Federated Storage.
WLCG Transfers monitoring EGI Technical Forum Madrid, 17 September 2013 Pablo Saiz on behalf of the Dashboard Team CERN IT/SDC.
Efi.uchicago.edu ci.uchicago.edu FAX status report Ilija Vukotic on behalf of the atlas-adc-federated-xrootd working group Computation and Enrico Fermi.
Efi.uchicago.edu ci.uchicago.edu Caching FAX accesses Ilija Vukotic ADC TIM - Chicago October 28, 2014.
Efi.uchicago.edu ci.uchicago.edu FAX splinter session Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Tier 1 / Tier 2 /
Federating Data in the ALICE Experiment
WLCG IPv6 deployment strategy
Introduction to Data Management in EGI
PanDA in a Federated Environment
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
FDR readiness & testing plan
Brookhaven National Laboratory Storage service Group Hironori Ito
Ákos Frohner EGEE'08 September 2008
Presentation transcript:

efi.uchicago.edu ci.uchicago.edu Data Federation Strategies for ATLAS using XRootD Ilija Vukotic On behalf of the ATLAS Collaboration Computation and Enrico Fermi Institutes University of Chicago Computing in High Energy and Nuclear Physics (CHEP 2013) Amsterdam, The Netherlands October 2013

efi.uchicago.edu ci.uchicago.edu What is FAX? Read only access Global namespace Currently 42 federated sites Regions covered: US, DE, UK, ES, and CERN Ilija Vukotic, CHEP October 2013, Amsterdam2 FAX (Federated ATLAS Xrootd) is a way to unify direct access to a diversity of storage services used by ATLAS

efi.uchicago.edu ci.uchicago.edu But Not only that! Initial use cases – Failover from stage-in problems with local storage – Gain access to more CPUs using WAN direct read access Allow brokering to Tier 2s with partial datasets Opportunistic resources without local ATLAS storage – Use as caching mechanism at sites to reduce local data management tasks Eliminate cataloging, consistency checking, deletion services WAN data access group formed in ATLAS to determine use cases & requirements on infrastructure Ilija Vukotic, CHEP October 2013, Amsterdam3

efi.uchicago.edu ci.uchicago.edu HOW it works Data can be asked for from any endpoint and redirector Data are transferred directly from server to user Searching for the file is fast, delivering is more important Ideally one should use the one with best connection Usually that’s the closest one Ilija Vukotic, CHEP October 2013, Amsterdam4 bnl NET2 AGLT2 MWT2 BNL OCHEP SWT2 SLAC XRD-CENTRAL GLRD ATLAS-XRD-EUATLAS-XRD-ASIA redirector endpoint

efi.uchicago.edu ci.uchicago.edu What are the ingredients? gLFN – global Logical Filename Protocol Storage systems at sites Redirectors Federation description Monitoring Storage service integration Applications Documentation Ilija Vukotic, CHEP October 2013, Amsterdam5 Simplicity for user requires a smart (therefore complex) system.

efi.uchicago.edu ci.uchicago.edu Global Logical Filenames In ATLAS data management based on DQ2 – Files are organized in DataSets, DataSets in DataContainers, exist in one or more SpaceTokens at one or more StorageElements and are registered in DQ2 – File catalog database LFC is an Oracle DB and contains mapping of logical file names to their physical path(s) – Each endpoint when asked for a file, queries LFC to find exact path of the file before delivering it or responding that file is not there – Proved to be large complication: multiple LFCs, authentication issues, scaling, latencies Now moving to new DDM system - RUCIO – Simple function will derive PFN from gLFN – Much faster, more reliable, easier to set up. Ilija Vukotic, CHEP October 2013, Amsterdam6

efi.uchicago.edu ci.uchicago.edu Access Protocol: XRootD The XROOTD project aims at giving high performance, scalable, fault tolerant access to data repositories of many kinds Widely used in High Energy Physics Supported by ROOT XRootD clients: xrdcp and xrdfs File path start with root://servername:[port]/… Can serve from any posix storage, dCache xrootd plugin implements protocol Ilija Vukotic, CHEP October 2013, Amsterdam7

efi.uchicago.edu ci.uchicago.edu A Diversity of Storage Systems ATLAS uses 80+ WLCG computing sites organized in Tiers Various storage technologies are used: dCache, DPM, Lustre, GPFS, Storm, XRootD and EOS Single source of deployment documentation: lasComputing/JoiningTheATLASFederationhttps://twiki.cern.ch/twiki/bin/view/At lasComputing/JoiningTheATLASFederation To ease support we have experts for different technologies To ease communications we have contact persons per national cloud Ilija Vukotic, CHEP October 2013, Amsterdam8

efi.uchicago.edu ci.uchicago.edu Storage Endpoints 41 sites 72% done All of USA, UK, DE, IT, RU, ES Most of FR Recently joined: Taiwan, PIC, IFAE Next: AU, CA, NL Ilija Vukotic, CHEP October 2013, Amsterdam9 T0T1T2D 1/16/1234/44 Coverage Continuous Status Monitoring

efi.uchicago.edu ci.uchicago.edu Redirector Network Lightweight and easily managed Ilija Vukotic, CHEP October 2013, Amsterdam10 GLRD EU ASIA US central UK DE FR IT ES RU

efi.uchicago.edu ci.uchicago.edu Describing the federation Both individual users and production/testing systems need to know access points, redirectors and their current status FAX topology recorded in the ATLAS Grid Info SystemATLAS Grid Info System Current status (and its history) are kept in Site Status Board (SSB)Site Status Board Ilija Vukotic, CHEP October 2013, Amsterdam11 Missing test dataset

efi.uchicago.edu ci.uchicago.edu Monitoring the infrastructure Ilija Vukotic, CHEP October 2013, Amsterdam12 xRootD Castor Collector at SLAC xRootD DPM Collector at CERN UDP Monitor xroot4j dCache xRootD posix xRootD MonaLisa ActiveMQ Consumers: Popularity DB SSB Dashboard Consumers: Popularity DB SSB Dashboard Monitor xroot4j 1 dCache Summary stream Detailed stream 1 dCache XRootD door

efi.uchicago.edu ci.uchicago.edu “Summary” Monitoring Stream Ilija Vukotic, CHEP October 2013, Amsterdam13 Lightweight XML – throughputs – connections – redirections – per site – per server Results stored in postgresql Shown in MonaLisa

efi.uchicago.edu ci.uchicago.edu “Detailed” Monitoring Stream Real time  actionable! File info – Path – Size – Open – Close – Reads/writes,… User info – Hostname – DN – Process ID Server info Application Info Ilija Vukotic, CHEP October 2013, Amsterdam14 Can see each currently open file Usually >5k files

efi.uchicago.edu ci.uchicago.edu Detailed Monitoring: WLCG dashboard Collects and keeps all the information Will make it easy to slice & dice data This is what shifters will look at when in production Ilija Vukotic, CHEP October 2013, Amsterdam15

efi.uchicago.edu ci.uchicago.edu Job Failover using FAX PanDA (distributed production and analysis system for ATLAS) sends a job only to a site having all the input data. In case that an input file can not be obtained after 2 tries, the file will be obtained through FAX in case it exists anywhere else. There are in average 2.8 copies of files from recent reprocessing in FAX so there is a large change of success. Ultimate success would be for a site to for example update its storage without first draining it of jobs. Ilija Vukotic, CHEP October 2013, Amsterdam16 Small number of jobs failing But these failures cost the most in terms of user’s turn- around time Small number of jobs failing But these failures cost the most in terms of user’s turn- around time

efi.uchicago.edu ci.uchicago.edu FAX usage in Job Brokering One can broker jobs to sites that don’t have all or part of input data and use FAX to read them directly. Beneficial in cases where A site has very often full queues as some specific data exist only there A site has free CPUs, good connectivity but not enough storage/data One can use external sources of CPU cycles (OSG, Amazon/Google cloud based queues,…) For this approach to be efficient, system has to “know” expected available bandwidth between a queue and all of the FAX endpoints We continuously measure this data and store it in Cost Matrix. We expect brokering to FAX to be functional in a few months. Ilija Vukotic, CHEP October 2013, Amsterdam17

efi.uchicago.edu ci.uchicago.edu Measures transfer rates (memory- to-memory) between 42 ANALY queues and each FAX endpoint Jobs submitted by HammerCloud Jobs send results to ActiveMQ, consumed by dedicated machine stored in SSB with other network performance measurements - perfSONAR and FileTransferService Ilija Vukotic, CHEP October 2013, Amsterdam18 HammerCloud FAX REST SSB FAX cost matrix SSB FAX cost matrix SSB Sonar view SSB Sonar view FTS PanDA job brokerage Cost matrix

efi.uchicago.edu ci.uchicago.edu FAX-based applications With dedicated or on-demand CPU, one can provide “MapReduce-like” web service applications Ongoing development – Event Picking Service Plucks small chunks of data from large numbers of data sets – Skim and Slim Service To help users go from multi-terabyte scale to multi-gigabyte scale A common grid use case A lot of optimizations possible if provided as a service Need to find balance between making it general and useful enough while limiting resource usage Ilija Vukotic, CHEP October 2013, Amsterdam19

efi.uchicago.edu ci.uchicago.edu Performance Metrics – Data Coverage >98%, >2.8 replicas – Number of users production, few tens of individual users – Percentage of successful jobs >99% from last tests – Total amount of data delivered ~2PB/week – Bandwidth usage Source – Ganglia plots – MonaLisa – FAX Dashboard – HC tests – CostMatrix tests – Special tests using dedicated resources Ilija Vukotic, CHEP October 2013, Amsterdam20

efi.uchicago.edu ci.uchicago.edu Performance HammerCloud based test Running real analysis code 100 concurrent jobs from each site Arrow width proportional to average event rate Ilija Vukotic, CHEP October 2013, Amsterdam21 WAN performance can be as good as LAN MWT2 SLAC AGLT2 BNL

efi.uchicago.edu ci.uchicago.edu Future Increase coverage, add redundancy, increase total bandwidth – Add the rest of sites Increase performance, reduce bandwidth needs – Caching – Cost matrix – smart FAX – Smart network - Bandwidth requests, QOS assurance Improve adoption rate – Presenting, teaching, preaching – New services Improve satisfaction – FAX tuning – Application tuning – New services Ilija Vukotic, CHEP October 2013, Amsterdam22

efi.uchicago.edu ci.uchicago.edu Conclusions Federation is functional First use cases implemented Usage is increasing Powerful tool that has to be used wisely Took more effort than initially expected – A lot of people (in different time zones) involved – Had to develop things multiple times for different storage technologies – Had to integrate it in our frameworks – Establishing effective monitoring as complex as establishing federation Ilija Vukotic, CHEP October 2013, Amsterdam23

efi.uchicago.edu ci.uchicago.edu Thank you! Questions? Ilija Vukotic, CHEP October 2013, Amsterdam24