OSG STORAGE OVERVIEW Tanya Levshina. Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients.

Slides:



Advertisements
Similar presentations
OSG Public Storage Project Summary Ted Hesselroth October 5, 2010 Fermilab.
Advertisements

 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
A. Sim, CRD, L B N L 1 Data Management Foundations Workshop, Mar. 3, 2009 Storage in OSG and BeStMan Alex Sim Scientific Data Management Research Group.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 7 Configuring File Services in Windows Server 2008.
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
16 th May 2006Alessandra Forti Storage Alessandra Forti Group seminar 16th May 2006.
SRM at Clemson Michael Fenn. What is a Storage Element? Provides grid-accessible storage space. Is accessible to applications running on OSG through either.
Introduction to OSG Storage Suchandra Thapa Computation Institute University of Chicago March 19, 20091GSAW 2009 Clemson.
OSG Services at Tier2 Centers Rob Gardner University of Chicago WLCG Tier2 Workshop CERN June 12-14, 2006.
A. Sim, CRD, L B N L 1 US CMS Workshop, Mar. 3, 2009 Berkeley Storage Manager (BeStMan) Alex Sim Scientific Data Management Research Group Computational.
CERN, 29 August 2006 Status Report Riccardo Zappi INFN-CNAF, Bologna.
OSG Middleware Roadmap Rob Gardner University of Chicago OSG / EGEE Operations Workshop CERN June 19-20, 2006.
The Disk Resource Manager A Berkeley SRM for Disk Cache Implementation and Testing Experience on the OSG Jorge Luis Rodriguez IBP/Grid Mini-Workshop Apr.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
Data management in grid. Comparative analysis of storage systems in WLCG.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
Tier 3 Data Management, Tier 3 Rucio Caches Doug Benjamin Duke University.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
G RID M IDDLEWARE AND S ECURITY Suchandra Thapa Computation Institute University of Chicago.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
OSG Storage Architectures Tuesday Afternoon Brian Bockelman, OSG Staff University of Nebraska-Lincoln.
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
The OSG and Grid Operations Center Rob Quick Open Science Grid Operations Center - Indiana University ATLAS Tier 2-Tier 3 Meeting Bloomington, Indiana.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Auditing Project Architecture VERY HIGH LEVEL Tanya Levshina.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
OSG Storage VDT Support and Troubleshooting Concerns Tanya Levshina.
VOX Project Tanya Levshina. 05/17/2004 VOX Project2 Presentation overview Introduction VOX Project VOMRS Concepts Roles Registration flow EDG VOMS Open.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
Tier 3 Support and the OSG US ATLAS Tier2/Tier3 Workshop at UChicago August 20, 2009 Marco Mambelli –
The GridPP DIRAC project DIRAC for non-LHC communities.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Mario Reale – GARR NetJobs: Network Monitoring Using Grid Jobs.
The VOMS and the SE in Tier2 Presenter: Sergey Dolgobrodov HEP Meeting Manchester, January 2009.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
OSG Status and Rob Gardner University of Chicago US ATLAS Tier2 Meeting Harvard University, August 17-18, 2006.
OSG STORAGE AND DATA MOVEMENT Tanya Levshina. Talk Outline  Data movement methods and limitation  Open Science Grid (OSG) Storage  Storage Resource.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
A. Sim, CRD, L B N L 1 Production Data Management Workshop, Mar. 3, 2009 BeStMan and Xrootd Alex Sim Scientific Data Management Research Group Computational.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
Open Science Grid Configuring RSV OSG Resource & Service Validation Thomas Wang Grid Operations Center (OSG-GOC) Indiana University.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
Grid Colombia Workshop with OSG Week 2 Startup Rob Gardner University of Chicago October 26, 2009.
BeStMan/DFS support in VDT OSG Site Administrators workshop Indianapolis August Tanya Levshina Fermilab.
KIT - University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association Xrootd SE deployment at GridKa WLCG.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
Federating Data in the ALICE Experiment
a brief summary for users
Jean-Philippe Baud, IT-GD, CERN November 2007
StoRM: a SRM solution for disk based storage systems
Berkeley Storage Manager (BeStMan)
Data Management cluster summary
Leigh Grundhoefer Indiana University
Summary of the dCache workshop
Presentation transcript:

OSG STORAGE OVERVIEW Tanya Levshina

Talk Outline  OSG Storage architecture  OSG Storage software  VDT cache  BeStMan  dCache  DFS:  SRM Clients  Auxiliary software  Statistics  OSG Storage Group  Summary 8/11/ OSG Site Admin Workshop

OSG Storage Architecture  Classic Storage Element  POSIX-mounted storage  Mounted and writable on the CE.  Readable from the worker nodes  Not-scalable under heavy load  High-performance FS is not cheap  Space management is not trivial  Storage Element  SRM endpoint  Provides GridFTP Load balancing  Transfers via GridFTP servers  May provide internal access protocols (xroot, Posix) Pictures from B. Bockelman’s presentation at OSS2010 8/11/ OSG Site Admin Workshop

Virtual Data Toolkit VDT provides:  A standard procedure for installation, configuration, services enabling, startup and shutdown  Simplified configuration scripts  All packages in one cache:  BeStMan  GridFTP  CA certificates, CRL installation, update  Log rotation scripts  Probes  Straightforward upgrade procedure 8/11/ OSG Site Admin Workshop

8/11/ OSG Site Admin Workshop

BeStMan-gateway  Generic SRM v2.2 load balancing frontend for GridFTP servers  Light-weight implementation of SRM v2.2 for POSIX file systems  srmPing,  srmLs  srmRm  srmMkdir  srmRmdir,  srmPrepareToPut (Status, PutDone),  srmPrepareToGet (Status, ReleaseFiles)  Designed to work with any Posix-like file systems  NFS, GPFS, GFS, Lustre, XrootdFS, HDFS  Doesn’t support queuing or disk space management  Hands-on installation will follow 8/11/ OSG Site Admin Workshop

BeStMan-gateway/Xrootd  Xrootd (developed at SLAC, contribution from CERN, others)  is designed to provide access POSIX-like via root framework (root://) Native commands (xrdcp,…)  Allows cluster globalization  Allows unix-like user/group authorization as well as X509 authentication.  Requires FUSE, XrootdFS to hook with BeStMan, GridFTP DSI plugin  Currently is used by many ATLAS and ALICE T2 sites, recommended for all Atlas T3  Can be installed from VDT (pacman) SE CE Worker node job SrmPut Submit root:// 8/11/ OSG Site Admin Workshop

SE BeStMan-gateway/HDFS  Hadoop DFS is developed in the Apache project.  Creates multiple replicas of data blocks  Distributes them on data nodes throughout a cluster  Consists of two major components: Namenode: central metadata server. Datanode: file servers for data  Allows replication  Runs on commodity hardware  unix-like user/group authorization, but no strict authentication  Requires FUSE to hook with BeStMan, GriFTP –HDFS plugin  BeStMan/HDFS and all auxiliary software can be installed from rpms (hands-on tutorial later today) Data Node Name Node fuse BeStMan GridFTP HDFS plugin CE Worker node job fuse NameServer SrmPut Submit 8/11/ OSG Site Admin Workshop

dCache  dCache is a distributed storage solution developed at DESY, Fermilab and NGDF  dCache supports requesting data from a tertiary storage system  Full SRM 2.2 implementation  nfs-mountable namespace  Multiple access protocols  Replica Manager  Role-based authorization  Information Provider  Probably, requires more administration then T3 may provide  Available from dcache.org and VDT with auxiliary software and installation/configuration script Picture from Ted Hesselroth’s (from presentation: “Installing and Using SRM-dCache” 8/11/ OSG Site Admin Workshop

SRM Clients  Available from VDT (OSG-Client, wn-client)  SRM-Fermi-Client commands  developed and maintained at Fermilab  access any Storage Element that complies with the SRM 1 or 2 specification  SRM-LBL-Client commands  developed at LBNL,  access any SRM v2.2 based storage components  LCG-utils is a suite of client tools for data movement written for the LHC Computing Grid.  based on the Grid File Access Library,  access any SRM v2.2 based storage components  May use logical file names and require a connection to a BDII-based catalog for some commands file copies and deletions, which take endpoints based on the SRM URL. 8/11/ OSG Site Admin Workshop

Gratia transfer probes  Included in BeStMan, dCache VDT Cache  Reports to Gratia Accounting System  Generates accounting information about file transfers, source, destination, size of the file and owner 8/11/ OSG Site Admin Workshop

RSV Storage probes  The Resource and Service Validation (RSV) provides monitoring infrastructure for an OSG site admin.  Client  Collector/Server  Periodic Availability Reports  Storage RSV probes:  Current probes: srm-ping, srm-copy  Coming soon: srmtester suite MyOSG 8/11/ OSG Site Admin Workshop

OSG SE Statistics These are the unofficial statistics based on BDII:  Number of sites providing Storage Elements: 49  Number of sites running dCache: 12  Number of sites running BeStMan-gateway: 37 HDFS 6 Xrootd 3 Lustre 3 REDDNet 1 All other sites: Local disk, NFS?  Number of sites reporting Gratia GridFTP Transfer Probes: 15 (daily transfer ~ files, 800 TB) 8/11/ OSG Site Admin Workshop

OSG Storage Group  Group members (all part time):  Ted Hesslroth (dcache, discovery tools)  Tanya Levshina (OSG Storage coordinator)  Abhishek Rana (hadoop)  Neha Sharma (support, dcache, probes,test suites)  Alex Sim (bestman develper and support)  Douglas Strain (rsv probes, xrootd, pigeon tools)  Packages certification  Test suites development, test stands  Auxiliary software development Gratia and RSV probes Discovery tools Pigeon tools (not in VDT yet)  Documentation  Support for site administrators GOC Tickets creation/monitoring Liaison to developers groups  Active mailing list: 8/11/ OSG Site Admin Workshop

Discovery and Pigeon Tools  Discovery tools provide a convenient way to discover storage elements and related information (surl, end path, available space) for a particular VO by queering BDII information.  Pigeon tools (created on top of Discovery tools) help a non-owner VO to debug site problems with Public Storage allocated for this VO  Runs periodically  Allows to see detailed errors  Allows to generate and monitor GOC ticket  Keeps archive  Will be available as RSV probes 8/11/ OSG Site Admin Workshop

Storage Documentation On OSG Twiki  Release Documentation:  Main Storage Page:  End User Guide  Site Admin Guide  Tier-3 specific documentation:  OSG Storage Group Meetings 8/11/ OSG Site Admin Workshop

Summary  There is plethora of available storage software solutions  Each solution has some pros and cons  Tier-3 coordinators are trying to come up with the most comprehensive solution that satisfies:  The needs of experiments  Hardware availability  Available efforts for installation, support and maintenance  VDT provides means to improve and simplify installation and configuration  OSG Storage group is ready to help! 8/11/ OSG Site Admin Workshop

Announcement  OSG Storage Forum  University of Chicago  September 21-22, 2010  General discussion of various storage solutions (new features, major improvements), scalability and performance. 8/11/2010OSG Site Admin Workshop 18