Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.

Slides:



Advertisements
Similar presentations
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
Advertisements

ATLAS T1/T2 Name Space Issue with Federated Storage Hironori Ito Brookhaven National Laboratory.
Xrootd Update OSG All Hands Meeting University of Nebraska March 19-23, 2012 Andrew Hanushevsky, SLAC
Duke and ANL ASC Tier 3 (stand alone Tier 3’s) Doug Benjamin Duke University.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
CERN IT Department CH-1211 Genève 23 Switzerland t Integrating Lemon Monitoring and Alarming System with the new CERN Agile Infrastructure.
Efi.uchicago.edu ci.uchicago.edu FAX status report Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago US ATLAS Computing Integration.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Integration Program Update Rob Gardner US ATLAS Tier 3 Workshop OSG All LIGO.
Multi-Tiered Storage with Xrootd at ATLAS Western Tier 2 Andrew Hanushevsky Wei Yang SLAC National Accelerator Laboratory 1CHEP2012, New York
Xrootd, XrootdFS and BeStMan Wei Yang US ATALS Tier 3 meeting, ANL 1.
Storage Wahid Bhimji DPM Collaboration : Tasks. Xrootd: Status; Using for Tier2 reading from “Tier3”; Server data mining.
OSG Area Coordinator’s Report: Workload Management April 20 th, 2011 Maxim Potekhin BNL
SLAC Experience on Bestman and Xrootd Storage Wei Yang Alex Sim US ATLAS Tier2/Tier3 meeting at Univ. of Chicago Aug 19-20,
Efi.uchicago.edu ci.uchicago.edu Towards FAX usability Rob Gardner, Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago US ATLAS.
Efi.uchicago.edu ci.uchicago.edu FAX meeting intro and news Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Federated Xrootd.
FAX Redirection Topology Wei Yang 7/16/121. Redirector hardware at CERN Redundant redirectors for EU, UK, DE, FR – Redundant (the “+” sign below) VMs.
Wahid, Sam, Alastair. Now installed on production storage Edinburgh: srm.glite.ecdf.ed.ac.uk  Local and global redir work (port open) e.g. root://srm.glite.ecdf.ed.ac.uk//atlas/dq2/mc12_8TeV/NTUP_SMWZ/e1242_a159_a165_r3549_p1067/mc1.
July-2008Fabrizio Furano - The Scalla suite and the Xrootd1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Redirector xrootd proxy mgr Redirector xrootd proxy mgr Xrd proxy data server N2N Xrd proxy data server N2N Global Redirector Client Backend Xrootd storage.
CERN IT Department CH-1211 Geneva 23 Switzerland GT WG on Storage Federations First introduction Fabrizio Furano
IN2P3Creating Federated Data Stores for the LHC Summary Andrew Hanushevsky SLAC National Accelerator Laboratory September 13-14, 2012 IN2P3,
Efi.uchicago.edu ci.uchicago.edu FAX status developments performance future Rob Gardner Yang Wei Andrew Hanushevsky Ilija Vukotic.
STATUS OF DCACHE N2N AND MONITORING REPORT I. CURRENT SITUATION xrootd4j is a part of dCache implemented in such a way that each change requires new dCache.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
02-June-2008Fabrizio Furano - Data access and Storage: new directions1.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Storage Federations and FAX (the ATLAS Federation) Wahid Bhimji University of Edinburgh.
Efi.uchicago.edu ci.uchicago.edu Status of the FAX federation Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Tier 1 /
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
XROOTD AND FEDERATED STORAGE MONITORING CURRENT STATUS AND ISSUES A.Petrosyan, D.Oleynik, J.Andreeva Creating federated data stores for the LHC CC-IN2P3,
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Efi.uchicago.edu ci.uchicago.edu Data Federation Strategies for ATLAS using XRootD Ilija Vukotic On behalf of the ATLAS Collaboration Computation and Enrico.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
Efi.uchicago.edu ci.uchicago.edu Ramping up FAX and WAN direct access Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
OSG Area Coordinator’s Report: Workload Management October 6 th, 2010 Maxim Potekhin BNL
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
11-June-2008Fabrizio Furano - Data access and Storage: new directions1.
Efi.uchicago.edu ci.uchicago.edu Storage federations, caches & WMS Rob Gardner Computation and Enrico Fermi Institutes University of Chicago BigPanDA Workshop.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Introduce Caching Technologies using Xrootd Wei Yang 10/28/14ATLAS TIM 2014 Univ. Chicago1.
Data Analysis w ith PROOF, PQ2, Condor Data Analysis w ith PROOF, PQ2, Condor Neng Xu, Wen Guan, Sau Lan Wu University of Wisconsin-Madison 30-October-09.
09-Apr-2008Fabrizio Furano - Scalla/xrootd status and features1.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
OSG STORAGE OVERVIEW Tanya Levshina. Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
DPM in FAX (ATLAS Federation) Wahid Bhimji University of Edinburgh As well as others in the UK, IT and Elsewhere.
Efi.uchicago.edu ci.uchicago.edu FAX splinter session Rob Gardner Computation and Enrico Fermi Institutes University of Chicago ATLAS Tier 1 / Tier 2 /
Efi.uchicago.edu ci.uchicago.edu Federating ATLAS storage using XrootD (FAX) Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
CMS data access Artem Trunov. CMS site roles Tier0 –Initial reconstruction –Archive RAW + REC from first reconstruction –Analysis, detector studies, etc.
Federating Data in the ALICE Experiment
a brief summary for users
Global Data Access – View from the Tier 2
Status of the SRM 2.2 MoU extension
dCache “Intro” a layperson perspective Frank Würthwein UCSD
dCache – protocol developments and plans
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Brookhaven National Laboratory Storage service Group Hironori Ito
Hironori Ito Brookhaven National Laboratory
Presentation transcript:

Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln

Again, why do we do this A data federation allows accessing data at other locations Opens many possibilities and use cases (need to understand/exploit pros and cons) See Rob Gardner’s slides at SW&C meeting last week: materialId=slides&confId=

Where are we in R&D & Deployment Identified site architectures Deployed a prototype federation with basic monitoring in the US Global name space issue addressed X509 implementing, deployed at SLAC Studying WAN reading performance using prun and TTreeCache Git repository to share configurations, knowledge Bi-weekly working group meetings 3

Site Architectures Six possible site architectures Work with backend storage systems: dCache, Xrootd, Posix Xrootd proxy based architecture works for all sites Deployed at AGLT2, BNL, UTA, OU, SLAC Overlapping xrootd on dCache Deployed at UC Xrootd with libXrdDcap.so (deprecated) Deployed at UIUC Working on direct access to dCache xrootd door Need N2N via authorization plugin 4

Deployment, Dashboard 5

Deployment, CMS-like monitoring 6

R&D, Technical Stuffs Global Name Space is addressed Global File Name to Storage File Name (N2N) translation ready N2N via a local LFC introduces small latency (3ms?) dq2-client 1.0 supports GFN X509 work in progress Implemented and working. Fixing a security hole Prun jobs run successfully against the federation Using prun -–pfnList to supply a list of input files in GFN Git repository at BNL 7

Next Steps Continue Deployment More sites with Xrootd 3.1.1, X509, etc. Can we do this in a month or two? Latency study and TTreeCache Probably need to get a sense of worst cases first Talk to experts Panda Analysis, for example: Have pilot replacing local missing files with GFN Take advantage of regional redirectors Local caching Or can we help ourselves? Note: lsm-get [-t token] [--size N] [--checksum csum] [--guid guid] source destination 8

Next Steps, cont’d Global redirector evaluation Reliability and performance explore the concept of regional redirector Monitoring Leaning toward DB-less detailed stream monitoring CMS-like monitoring: Know what is happening, needs aggregation and history T3mon Bug fixes Extreme copy mode in Xrdcp Patch X509 security issue 9

Backup Slides 10

Redirector xrootd proxy mgr Redirector xrootd proxy mgr Xrd proxy data server N2N Xrd proxy data server N2N Global Redirector Client Backend Xrootd storage Backend Xrootd storage 0. Join the Federation 1. Open (GFN) 2. Open (GFN) 3. Open/Read (GFN) 4. Read (PFN)Fetched data data Export Xrootd storage via Xrootd Proxy cluster GFN: global file name PFN: physical (storage) file name N2N: translation GFN to PFN cms.dfs directive: see cms.dfs : all data servers are equal

Redirector Xrootd mgr Redirector Xrootd mgr Regluar Xrd data server N2N Regular Xrd data server N2N Global Redirector Client Backend Posix storage Backend Posix storage 0. Join the Federation 1. Open (GFN) 2. Open (GFN) 3. Open/Read (GFN) 4. Read (PFN)Fetched data data Export Posix storage via regular Xrootd cluster GFN: global file name PFN: physical (storage) file name N2N: translation GFN to PFN Posix file system: GPFN, Luster, XrootdFS cms.dfs : all data servers are equal Posix File system

Redirector Xrootd mgr Redirector Xrootd mgr Regular Xrd data server with N2N Regular Xrd data server with N2N Global Redirector Client Backend dCache dCap doors Backend dCache dCap doors 0. Join the Federation 1. Open (GFN) 2. Open (GFN) 3. Open/Read (GFN) 4. Read (PFN)Fetched data data Export dCache dCap doors via regular Xrootd cluster (deprecated) GFN: global file name PFN: physical (storage) file name N2N: translation GFN to PFN dCap plugin: a Xrootd OFS plug-in moudel using dCap protocol cms.dfs : all data servers are equal dCap plug-in

Redirector xrootd proxy mgr Redirector xrootd proxy mgr Xrd proxy data server N2N Xrd proxy data server N2N Global Redirector Client Backend dCache Xrootd doors Backend dCache Xrootd doors 0. Join the Federation 1. Open (GFN) 2. Open (GFN) 3. Open/Read (GFN) 4. Read (PFN)Fetched data data Export dCache Xrootd doors via Xrootd Proxy cluster, 1 GFN: global file name PFN: physical (storage) file name N2N: translation GFN to PFN cms.dfs : all data servers are equal

Regular xrootd data server with Xrootd.redirect … Global Redirector Client 0. Join the Federation 1. Open (GFN) 2. Open (GFN) data Export dCache Xrootd doors via Xrootd Proxy cluster, 2 GFN: global file name PFN: physical (storage) file name N2N: translation GFN to PFN For Xrootd.redirect directive, see: dCache pool node with Auth plug-in for N2N dCache pool node with Auth plug-in for N2N dCache pool node with Auth plug-in for N2N dCache pool node with Auth plug-in for N2N dCache xrootd door Auth plug-in for N2N dCache xrootd door Auth plug-in for N2N 3. Open (GFN) 4. Open/Read (GFN)

Redirector Xrootd mgr Redirector Xrootd mgr Global Redirector Client 0. Join the Federation 1. Open (GFN) 2. Open (GFN) 3. Open/Read (GFN) data Overlapping Xrootd cluster on top of dCache GFN: global file name PFN: physical (storage) file name N2N: translation GFN to PFN GFN name space : native xrootd name space layout according to GFN, with symlinks pointing to actual file on the same node (some CMS sites do this) dCache pool node Native Xrootd data server GFN name space (symlinks) or N2N dCache pool node Native Xrootd data server GFN name space (symlinks) or N2N