SRM Space Tokens Scalla/xrootd Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-May-08

Slides:



Advertisements
Similar presentations
Potential Data Access Architectures using xrootd OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Advertisements

Andrew Hanushevsky7-Feb Andrew Hanushevsky Stanford Linear Accelerator Center Produced under contract DE-AC03-76SF00515 between Stanford University.
Lecture 10: The FAT, VFAT, and NTFS Filesystems 6/17/2003 CSCE 590 Summer 2003.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
 AFS name space managed as mini-file systems (aka Volumes)  AFS Mount Points are the equivalent of NTFS Junctions and DFS Referrals which are stored.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Module 10 Configuring and Managing Storage Technologies.
Storage and NT File System INFO333 – Lecture Mariusz Nowostawski Noria Foukia.
Scalla Back Through The Future Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 8-April-10
Hands-On Microsoft Windows Server 2008
XRootD Roadmap To Start The Second Decade Root Workshop Saas-Fee March 11-14, 2013 Andrew Hanushevsky, SLAC
Xrootd Authentication & Authorization Andrew Hanushevsky Stanford Linear Accelerator Center 6-June-06.
File System Access (XRootd) Andrew Hanushevsky Stanford Linear Accelerator Center 13-Jan-03.
The Next Generation Root File Server Andrew Hanushevsky Stanford Linear Accelerator Center 27-September-2004
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 19-August-2009 Atlas Tier 2/3 Meeting
Xrootd Demonstrator Infrastructure OSG All Hands Meeting Harvard University March 7-11, 2011 Andrew Hanushevsky, SLAC
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 19-May-09 ANL Tier3(g,w) Meeting.
Scalla/xrootd Andrew Hanushevsky SLAC National Accelerator Laboratory Stanford University 29-October-09 ATLAS Tier 3 Meeting at ANL
Guide to Linux Installation and Administration, 2e1 Chapter 2 Planning Your System.
Xrootd, XrootdFS and BeStMan Wei Yang US ATALS Tier 3 meeting, ANL 1.
Scalla/xrootd Introduction Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 6-April-09 ATLAS Western Tier 2 User’s Forum.
July-2008Fabrizio Furano - The Scalla suite and the Xrootd1 cmsd xrootd cmsd xrootd cmsd xrootd cmsd xrootd Client Client A small 2-level cluster. Can.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Xrootd Monitoring Atlas Software Week CERN November 27 – December 3, 2010 Andrew Hanushevsky, SLAC.
Are SE Architectures Ready For LHC? Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 3-November-08 ACAT Workshop.
July-2008Fabrizio Furano - The Scalla suite and the Xrootd1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Scalla Developer Summary xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory CERN Workshop 10-November-08
Redirector xrootd proxy mgr Redirector xrootd proxy mgr Xrd proxy data server N2N Xrd proxy data server N2N Global Redirector Client Backend Xrootd storage.
Status & Plan of the Xrootd Federation Wei Yang 13/19/12 US ATLAS Computing Facility Meeting at 2012 OSG AHM, University of Nebraska, Lincoln.
The Replica Location Service The Globus Project™ And The DataGrid Project Copyright (c) 2002 University of Chicago and The University of Southern California.
CERN IT Department CH-1211 Genève 23 Switzerland Internet Services Xrootd explained
02-June-2008Fabrizio Furano - Data access and Storage: new directions1.
Accelerating Debugging In A Highly Distributed Environment CHEP 2015 OIST Okinawa, Japan April 28, 2015 Andrew Hanushevsky, SLAC
Performance and Scalability of xrootd Andrew Hanushevsky (SLAC), Wilko Kroeger (SLAC), Bill Weeks (SLAC), Fabrizio Furano (INFN/Padova), Gerardo Ganis.
ATLAS XRootd Demonstrator Doug Benjamin Duke University On behalf of ATLAS.
Scalla/xrootd Andrew Hanushevsky, SLAC SLAC National Accelerator Laboratory Stanford University 08-June-10 ANL Tier3 Meeting.
Scalla Advancements xrootd /cmsd (f.k.a. olbd) Fabrizio Furano CERN – IT/PSS Andrew Hanushevsky Stanford Linear Accelerator Center US Atlas Tier 2/3 Workshop.
Module 12: Configuring and Managing Storage Technologies
Scalla Authorization xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory CERN Seminar 10-November-08
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
Xrootd Proxy Service Andrew Hanushevsky Heinz Stockinger Stanford Linear Accelerator Center SAG September-04
Scalla In’s & Out’s xrootdcmsd xrootd /cmsd Andrew Hanushevsky SLAC National Accelerator Laboratory OSG Administrator’s Work Shop Stanford University/SLAC.
Scalla As a Full-Fledged LHC Grid SE Wei Yang, SLAC Andrew Hanushevsky, SLAC Alex Sims, LBNL Fabrizio Furano, CERN SLAC National Accelerator Laboratory.
Scalla + Castor2 Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-March-07 Root Workshop Castor2/xrootd.
Federated Data Stores Volume, Velocity & Variety Future of Big Data Management Workshop Imperial College London June 27-28, 2013 Andrew Hanushevsky, SLAC.
1 Xrootd-SRM Andy Hanushevsky, SLAC Alex Romosan, LBNL August, 2006.
11-June-2008Fabrizio Furano - Data access and Storage: new directions1.
Bestman & Xrootd Storage System at SLAC Wei Yang Andy Hanushevsky Alex Sim Junmin Gu.
09-Apr-2008Fabrizio Furano - Scalla/xrootd status and features1.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Scalla Update Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 25-June-2007 HPDC DMG Workshop
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
Storage Architecture for Tier 2 Sites Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 8-May-07 INFN Tier 2 Workshop
Federating Data in the ALICE Experiment
StoRM: a SRM solution for disk based storage systems
Xrootd explained Cooperation among ALICE SEs
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Installing OS.
SLAC National Accelerator Laboratory
StoRM Architecture and Daemons
Data access and Storage
Ákos Frohner EGEE'08 September 2008
Scalla/XRootd Advancements
Large Scale Test of a storage solution based on an Industry Standard
Storage Virtualization
INFNGRID Workshop – Bari, Italy, October 2004
Summary of the dCache workshop
Presentation transcript:

SRM Space Tokens Scalla/xrootd Andrew Hanushevsky Stanford Linear Accelerator Center Stanford University 27-May-08

27-May-082: Outline Introduction SRM Static Space Token Refresher Space Tokens in the Scalla Architecture Disk partitions as a space token paradigm How it was done Space usage and quotas by space token New Stuff Proxies unlimited Announcements Conclusion & Future Outlook

27-May-083: SRM Static Space Tokens Encapsulate fixed space characteristics Type of space E.g., Permanence, performance, etc. Imply a specific quota Using a particular arbitrary name E.g., data, dq2, mc, etc Typically used to create new files Think of it as a space profile

27-May-084: SRM Space Tokens & Paths Static Space Tokens may be redundant True if exclusive correspondence exists TokenPath dq2/atlas/dq2/…. mc/atlas/mc/…. Space tokens useful for overlapping namespaces Makes space token utility non-obvious But I digress, so let’s move on….

27-May-085: Space Tokens & xrootd Space attribute concept already part of xrootd Embodied by notion of cgroup A cgroup is a logical name for one or more file systems Implemented in the standard oss plug-in Used by default libXrdOfs.so plug-in The real work was to support SRM concepts Largely in the area of virtual quotas Opportunity to greatly improve the implementation

27-May-086: lfn2pfn prefix encoding Storage System (oss, drm/srm, etc) authentication (gsi, krb5, etc) Clustering (cmsd) authorization (name based) File System (ofs, sfs, alice, etc) Protocol (1 of n) (xrootd) Where Do Space Tokens Apply? Protocol Driver (XRD) SRM Static Space Tokens Implemented Here

27-May-087: Partitions as a Space Token Paradigm Disk partitions map well to SRM space tokens A set of partitions embody a set of space attributes Performance, quota, etc. A static space token defines a set of space attributes Partitions and static space tokens are interchangeable xrootd already supports multiple partitions Real as well as virtual partitions Can leverage this support for SRM space token support So, on to xrootd partition management

27-May-088: Partition Architecture N real partitions can be aggregated Each aggregation is called a virtual partition Uniform name space across all partitions Real partitions are space load balanced Reduces the granularity of failure Implemented via symlinks from a name space Name space itself resides in a real partition symlinkssymlinks Name Space Disk Space

27-May-089: Virtual vs Real Partitions Simple two step process Define your real partitions (one or more) These are file system mount-points Map virtual partitions on top of real ones Virtual partitions can share real partitions By convention, virtual partitions equal static token names Yields implicit SRM space token support oss.cache public /store1 xa oss.cache dq2 /store2 xa oss.cache data /store3 xa Virtual Partitions Real Partitions What’s this?

27-May-0810: xa Introducing xa Partitions Original oss partition architecture was limited Simplistic symlink target names Constrained file path length to 255 or less Could not automatically track assigned space tokens xa The xa option introduced for SRM support Supports paths up to 1024 characters Automatically tracks assigned space token Tracks usage for real and virtual partitions Both supported for backward compatibility xa The xa version is now preferred in all cases

27-May-0811: Partition Aggregation /store1/public /atlas/mydatafile /store3/data symlink Name Space Partition Holds Name Space Can also optionally hold data files Independent Filesystems Mount Points Symlink conventions simplify audits & name space recovery /store1 /store2 /store3 /store2/dq2/00/ % /store2/dq2/00/ %.pfn oss.cache public /store1 xa oss.cache dq2 /store2 xa oss.cache data /store3 xa /store2/data # oss.cache data /store2 xa Virtual partitions dq2 & data now share same Real partition

27-May-0812: Partition Selection Partitions selected by virtual partition name Configuration file: New files “cgi-tagged” with virtual partition name root://host:1094//atlas/mydatafile?cgroup=dq2 The default is “public” File allocated in a real partition associated with the named virtual partition By convention, the name is the SRM space token name oss.cache public /store1 xa oss.cache dq2 /store2 xa oss.cache data /store3 xa

27-May-0813: Real vs Virtual Partitions A real partition represents a hard quota Non-overlapping virtual partitions are real Simple and very effective Typically not efficiently utilized Shared real partitions Overlapping virtual partitions are virtual Provide better space utilization, but… Need usage tracking and quota management

27-May-0814: Partition Usage Tracking Usage is tracked by partition Automatic for real partitions Configurable for virtual partitions oss.usage {nolog | log dirpath} As Virtual Partition  SRM Space Token Usage is also automatically tracked by space token POSIX getxattr() returns usage information See Linux man page

27-May-0815: Partition Quota Management Quotas applied by partition Automatic for real partitions Configurable for virtual partitions oss.usage quotafile filepath POSIX getxattr() returns quota information Used by Fuse/xrootd to enforce quotas Fuse has view of the complete cluster Using the cluster name space daemon

27-May-0816: The Quota File Lists quota for each virtual partition Hence, also a quota for each static space token Simple multi-line format vpname nnnn[k | m | g | t]\n Re-read whenever it changes Useful only in the context of the cnsd xrootd Quotas need to apply to the whole cluster Investigating native integration with the redirector Currently, only FUSE/xrootd enforces quotas

27-May-0817: Other Considerations Files cannot be easily reassigned space tokens Must manually “move” file across partitions Partitions 1-to-1 correspondence with space tokens Can always get original space token name Use file-specific getxattr() call Quotas for virtual partitions are “soft” Time causality prevents hard limit Use real partitions if hard limit needed

27-May-0818: Proxies Unlimited Classic Proxy Server Restricted to a very specific role cmsd xrootd xroot cluster Proxy plugin proxy node xrootd FUSE Fuse/xrootd xrootd effective proxy node Introducing F FF FUSE as a proxy All cluster features available We are still investigating this exciting concept

27-May-0819: FUSE FUSE Proxy Transfer Rates SLAC#Streams BNL MB/sCERN MB/s scp gridftp xrdcp xrdcp xrdcp xrdcp We don’t have any explanations yet. We need to do more tests. (especially with multi-stream gridftp) FUSE Fuse/xrootd xrootd effective proxy node

27-May-0820: Secure Data Sharing Achievable FUSE Fuse/xrootd xrootd effective proxy node FUSE Fuse/xrootd xrootd effective proxy node FUSE Fuse/xrootd xrootd effective proxy node FUSE Fuse/xrootd xrootd effective proxy node BNL UTA WISC SLAC Data everywhere US Atlas Redirectors

27-May-0821: Announcements! The CERN-based Scalla web page is online! Scalla CVS repository is going public! Will be located in afs with unrestricted read access Planning on providing web access

27-May-0822: Conclusion & Future Outlook Scalla/xrootd is living up to our expectations Relatively easy to add SRM space token support More improvement are in the pipeline Kernel memory direct data transfer (a.k.a. sendfile) Significant reduction in CPU/Memory usage DSS Directed Support Services (DSS) Architecture A take-off on Multics special files Currently supporting Just-In-Time Alice Data Analysis Bandwidth management during on-demand data transfers Framework for simple intra-cluster resource management

27-May-0823: Acknowledgements Current software Collaborators Andy Hanushevsky, Fabrizio Furano Root: Fons Rademakers, Gerri Ganis (security), Bertrand Bellenot (windows) Alice: Derek Feichtinger, Andreas Peters, Guenter Kickinger STAR/BNL: Pavel Jackl, Jerome Lauret SLAC: Jacek Becla, Tofigh Azemoon, Wilko Kroeger Operational collaborators BNL, CERN, CNAF, FZK, INFN, IN2P3, GSI, RAL, SLAC SLAC Funding US Department of Energy Contract DE-AC02-76SF00515 with Stanford University