Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland www.cern.ch/i t DBCF GT Standard Protocols in DPM Ricardo Rocha.

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR Status Alberto Pace.
The GridSite Toolbar Shiv Kaushal The University of Manchester All Hands Meeting 2006.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT dpm-xrootd v3 Creating Federated Data Stores for the LHC David.
Naming Computer Engineering Department Distributed Systems Course Asst. Prof. Dr. Ahmet Sayar Kocaeli University - Fall 2014.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Simplifying Configuration Ricardo Rocha ( on behalf of the LCGDM.
CERN IT Department CH-1211 Genève 23 Switzerland t Integrating Lemon Monitoring and Alarming System with the new CERN Agile Infrastructure.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Almaden Rice University Nache: Design and Implementation of a Caching Proxy for NFSv4 Ajay Gulati, Rice University Manoj Naik, IBM Almaden Renu Tewari,
10 May 2007 HTTP - - User data via HTTP(S) Andrew McNab University of Manchester.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
A comparison of distributed data storage middleware for HPC, GRID and Cloud Mikhail Goldshtein 1, Andrey Sozykin 1, Grigory Masich 2 and Valeria Gribova.
CERN IT Department CH-1211 Genève 23 Switzerland t Plans and Architectural Options for Physics Data Analysis at CERN D. Duellmann, A. Pace.
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
Ricardo Rocha ( on behalf of the DPM team ) Standards, Status and Plans.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Future home directories at CERN
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM Collaboration Motivation and proposal Oliver Keeble CERN On.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Interfaces to Grid Storage DPM and LFC Update Ricardo.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Operating Systems & Information Services CERN IT Department CH-1211 Geneva 23 Switzerland t OIS Drupal at CERN Juraj Sucik Jarosław Polok.
Andrew McNab - HTTP/HTTPS extensions HTTP/HTTPS as Grid data transport 6 March 2003 Andrew McNab, University of Manchester
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
Distributed File Systems Questions answered in this lecture: Why are distributed file systems useful? What is difficult about distributed file systems?
SESEC Storage Element (In)Security hepsysman, RAL 0-1 July 2009 Jens Jensen.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
EMI is partially funded by the European Commission under Grant Agreement RI Roadmap & Future Work Ricardo Rocha ( on behalf of the DPM team )
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Cluman: Advanced Cluster Management for Large-scale Infrastructures.
Security recommendations DPM Jean-Philippe Baud CERN/IT.
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Solutions for WAN data access: xrootd and NFSv4.1 Andrea Sciabà.
EMI is partially funded by the European Commission under Grant Agreement RI DPM in EMI-II HTTP and NFS interfaces Oliver Keeble On behalf of DPM.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
Dynamic Federation of Grid and Cloud Storage Fabrizio Furano, Oliver Keeble, Laurence Field Speaker: Fabrizio Furano.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
CERN IT-Storage Strategy Outlook Alberto Pace, Luca Mascetti, Julien Leduc
EGEE Data Management Services
Onedata Eventually Consistent Virtual Filesystem for Multi-Cloud Infrastructures Michał Orzechowski (CYFRONET AGH)
Jean-Philippe Baud, IT-GD, CERN November 2007
Dynamic Storage Federation based on open protocols
Ricardo Rocha ( on behalf of the DPM team )
DPM Installation Configuration
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
Onedata Eventually Consistent Virtual Filesystem for Multi-Cloud Infrastructures Michał Orzechowski (CYFRONET AGH)
Introduction to Data Management in EGI
GFAL 2.0 Devresse Adrien CERN lcgutil team
CTA: CERN Tape Archive Overview and architecture
Data Management cluster summary
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha

Grid Technology DPM Main Goals Provide lightweight “grid enabled” storage Manage space on distributed disk servers Manage a hierarchical namespace Expose interfaces for 3 use cases: –Space management (socket, SRM1.1, SRM2.2) –Remote data access (gridFTP, HTTP/HTTPS) –POSIX like access (rfio, xroot)

Grid Technology DPM Further Details Written mostly in C C/C++ API, Python API, Perl API, CLI Statistics from GStat – –Over 200 grid sites use DPM Largest deployment: 1.5PB More Information –

Grid Technology Why Standards

Grid Technology Standards help with… Library dependency issues Requirement of user interfaces (UIs) –Entry points to the grid –Maintained by experts Very hard to use “standard” distributions –Even transition from SLC4 to 5 is problematic Validation takes a long time

Grid Technology Standards improve… Accessibility –Not limiting access to OS X version Y with library Z Validation –Using common validation and test tools Stability –Evolution discussed in a wide group Ease of implementation –Sharing of experiences, common code base No vendor lock-in …

Grid Technology Use Case: Space Management SRM

Grid Technology Storage Resource Manager (SRM) An interface to manage storage resources –Supported by all grid storage backends DPM supports all SRM versions in use –1.1 / 2.2 Stable, in production for several years Current focus: provide pure SSL access –Ongoing effort in the context of EMI (together with dCache, BeStMan, STORM, …) –Replacing httpg with simple https –Requires alternative for credential delegation

Grid Technology Use Case: Remote Data Access HTTP / WebDAV

Grid Technology HTTP / WebDAV DPM supports file access via HTTP –Currently only as a transfer protocol –Implemented as an apache module or cgi Firewall friendly Extensive client availability Work in progress: –Adding support for WebDAV (RFC 2518): locking, namespace management, metadata on files Currently not so interesting within HEP, but very popular among other communities –Improve the HTTP implementation (benefit from upcoming nameserver code refactoring) Major focus once NFS 4.1 is stable

Grid Technology Use Case: POSIX Data Access NFS 4.1

Grid Technology NFS 4.1 Very good match for our needs Strong authentication / authorization –KRB5 (ongoing work for X509), Basic permissions & ACLs Performance –Via bulk operations (few round trips), Sessions, Delegations –Even more relevant for WAN access Parallel / Clustered data access –Parallel NFS (pNFS), multiple layout types (file, object, block) Multi server namespace –Spawning multiple domains, Client redirection, Alternative replica selections Free Clients (Linux, Solaris, Windows) –Supported by major industry vendors, Free caching Free system administration and monitoring tools

Grid Technology DPM NFS 4.1 Details (client and server): Most work is on the Head Node (Metadata Server) Possibility to reuse existing open source code Looked into two alternatives –sPNFS: a kernel based implementation Working prototype, but kernel programming is hard –User space daemon (Ganesha) Mainly developed by CEA in France, in production for several years First DPM enabled release deployed internally, in validation phase For the moment we focus on the second option (but did not completely drop the first)

Grid Technology DPM NFS 4.1 DPM HEAD NODE DATA SERVER DATA SERVER CONTROL PROTOCOL (undefined) NFS4.1/pNFS ProtocolStorage Access Protocol (File Layout) DATA SERVER DATA SERVER … … Callbacks CLIENT Standard Distribution Kernel

Grid Technology DPM NFS 4.1: Status & Roadmap Prototype ready and deployed internally –Includes strong security –Automated build and testing environment –Finishing functionality validation Next step is to extend testing (starting now) –Increase the number of sites involved –Include a DPM instance in the CERN NFS 4.1 performance testing effort Wide deployment first half 2011 Further work: –Multi server namespace : global grid namespace and client redirection

Grid Technology DPM NFS 4.1: Multi Server Namespace Namespace spawning multiple domains Servers redirect clients when data is not local –Redirection is the key word here Can also be used to provide clients with alternative locations (transparent failover on server overload, unavailability) CLIENT NFS SITE A NFS SITE B 1 2 open (/grid/siteB/myFile) NFS4ERR_MOVED ( fs_locations )

Grid Technology DPM NFS 4.1: Extra goodies We can benefit from using a standard Existing validation / testing suites –Example: PyNFS Existing monitoring tools –Example: collectl Existing debugging tools –Example: wireshark / tshark #./testserver.py $DPMNFS_HOST:/grid/dpm/cern.ch -v dir # collectl -snf

Grid Technology DPM NFS 4.1: Extra goodies We can benefit from using a standard Existing validation / testing suites –Example: PyNFS Existing monitoring tools –Example: collectl Existing debugging tools –Example: wireshark / tshark #./testserver.py $DPMNFS_HOST:/grid/dpm/cern.ch -v dir # collectl -snf # tshark -R "nfs" -V Network File System, Ops(4): PUTFH LOOKUP GETFH GETATTR [Program Version: 4] [V4 Procedure: COMPOUND (1)] Tag: length: 0 contents: minorversion: 0 Operations (count: 4) Opcode: PUTFH (22) filehandle length: 112 [hash: 0x311d234b] decode type as: unknown filehandle: D Opcode: LOOKUP (15) Filename: cern.ch length: 7 contents: cern.ch …

Grid Technology Conclusion With HTTP/WebDAV and NFS4.1, DPM provides standard based solutions for all its use cases Benefits exist for both clients and system administrators (and even developers) DPM will continue its work on improving the status of grid data storage and access