Ricardo Rocha ( on behalf of the DPM team ) Standards, Status and Plans.

Slides:



Advertisements
Similar presentations
PRESENTATION TITLE GOES HERE Introduction to NFS v4 and pNFS David Black, SNIA Technical Council, EMC slides by Alan Yoder, NetApp with thanks to Michael.
Advertisements

30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
DPM Basics and its status and plans Wahid Bhimji University of Edinburgh GridPP Storage Workshop – Apr 2010 Apr-101Wahid Bhimji – DPM.
DPM Status & Roadmap Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
HDFS and S3 plugins Andrea Manzi Martin Hellmich 13/12/2013.
EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
10 May 2007 HTTP - - User data via HTTP(S) Andrew McNab University of Manchester.
Sensitivity of Cluster File System Access to I/O Server Selection A. Apon, P. Wolinski, and G. Amerson University of Arkansas.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
New DM Clients and retirement plans for old ones Alejandro Alvarez Ayllon on behalf of the DM Clients developers IT/SDC 11/12/2013.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
1 Week #10Business Continuity Backing Up Data Configuring Shadow Copies Providing Server and Service Availability.
EMI 1 Release The EMI 1 (Kebnekaise) release features for the first time a complete and consolidated set of middleware components from ARC, dCache, gLite.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
RAL Site Report Castor Face-to-Face meeting September 2014 Rob Appleyard, Shaun de Witt, Juan Sierra.
Alejandro Alvarez Ayllon on behalf of the LCGDM developer team IT/SDC 13/12/2013 DAV support in DPM.
GLUE 2 Open Issues in Storage Information Providers 16 th May 2014.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
DotSlash – or how to deal with 15 minutes of fame Weibin Zhao Henning Schulzrinne Columbia University CATT/WICAT Annual Research Review November 14, 2003.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
EMI INFSO-RI Overview of the EMI development objectives Balázs Kónya (Lund University) EMI Technical Director EMI All Hands Prague, 23rd November.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Interfaces to Grid Storage DPM and LFC Update Ricardo.
CERN IT Department CH-1211 Geneva 23 Switzerland GT HTTP solutions for data access, transfer, federation Fabrizio Furano (presenter) on.
First Look at the New NFSv4.1 Based dCache Art Kreymer, Stephan Lammel, Margaret Votava, and Michael Wang for the CD-REX Department CD Scientific Computing.
Storage Interfaces Introduction Wahid Bhimji University of Edinburgh Based on previous discussions with Working Group: (Brian Bockelman, Simone Campana,
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Andrew McNab - HTTP/HTTPS extensions HTTP/HTTPS as Grid data transport 6 March 2003 Andrew McNab, University of Manchester
Storage Interfaces and Access pre-GDB Wahid Bhimji University of Edinburgh On behalf of all those who participated.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Overview of DMLite Ricardo Rocha ( on behalf of the LCGDM team.
GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
EMI is partially funded by the European Commission under Grant Agreement RI Roadmap & Future Work Ricardo Rocha ( on behalf of the DPM team )
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
EMI INFSO-RI Patrick Fuhrmann EMI Data area leader At the EGI Technical Forum 2011, in Lyon EMI-Data The second year.
CRISP WP18, High-speed data recording Krzysztof Wrona, European XFEL PSI, 18 March 2013.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Security recommendations DPM Jean-Philippe Baud CERN/IT.
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
FILE TRANSFER SPEEDS OVER HTTP AND FTP Yibiao Li 06/01/2009 Christmas Meeting 2008/09.
EMI is partially funded by the European Commission under Grant Agreement RI DPM in EMI-II HTTP and NFS interfaces Oliver Keeble On behalf of DPM.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
EGEE Data Management Services
Jean-Philippe Baud, IT-GD, CERN November 2007
File Syncing Technology Advancement in Seafile -- Drive Client and Real-time Backup Server Johnathan Xu CTO, Seafile Ltd.
Ricardo Rocha ( on behalf of the DPM team )
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
dCache “Intro” a layperson perspective Frank Würthwein UCSD
Gfal/lcg-util -> Gfal2/gfal2-util
Introduction to Data Management in EGI
GFAL 2.0 Devresse Adrien CERN lcgutil team
WLCG Storage planning Andrea Sciabà
Data Management cluster summary
Presentation transcript:

Ricardo Rocha ( on behalf of the DPM team ) Standards, Status and Plans

EMI INFSO-RI CLIENT DPNSDPMSRMHTTPNFS GRIDFTP RFIOHTTPNFSXROOT HEAD NODE DISK NODE(s) FILE METADATA OPS FILE ACCESS OPS RFIO HTTP NFS XROOT DPM Overview

EMI INFSO-RI DPM Core 1.8.2, Testing, Roadmap

EMI INFSO-RI DPM – Highlights Improved scalability of all frontend daemons – Especially with many concurrent clients – By having a configurable number of threads Fast/Slow in case of the dpm daemon Faster DPM drain – Disk server retirement, replacement, … Better balancing of data among disk nodes – By assigning different weights to each filesystem Log to syslog GLUE2 support

EMI INFSO-RI DPM Core : Extended Testing Activity Thanks to ShuTing for the plots ( preliminary results ) HC using GridFTP HC using RFIO Cluster at ASGC (thanks!) 1000 Cores Regular Hammercloud Tests

EMI INFSO-RI DPM Core – Roadmap Package consolidation: EPEL compliance Fixes in multi-threaded clients Replace httpg with https on the SRM Improve dpm-replicate (dirs and FSs) GUIDs in DPM Synchronous GET requests Reports on usage information Quotas Accounting metrics HOT file replication

EMI INFSO-RI DPM Beta Components HTTP/DAV, NFS, Nagios, Puppet, Perfsuite, Catalog Sync, Contrib Tools

EMI INFSO-RI DPM Beta: HTTP / DAV Overview, Performance

EMI INFSO-RI DPM HTTP / DAV: Overview CLIENT LFC DPM HEAD DPM DISK GET GET / PUT REDIRECT DATA

EMI INFSO-RI DPM HTTP / DAV: Overview

EMI INFSO-RI HTTP : Client Support curlbrowser OSAny GUINOYES CLIYESNO X509YES ProxiesYESOnly IE so far RedirectYES PUTYESNO Recommendation: browser/curl for GET, curl for PUT Chrome Issue 9056 submitted for proxy support

EMI INFSO-RI DAV : Client Support TrailMixCadaverDavlibShared Folder DavFS2NautilusDolphin OSFirefox < 4*nixMac OS XWindows*nixGnomeKDE GUIYESNOYES N/AYES CLINOYESNO N/ANO X509YES NOYES NO Proxies?NO YESNO RedirectYESNOYESNot PUTNO YES Updated analysis based on initial one from dCache Recommendation: Cadaver for *nix, Windows explorer

EMI INFSO-RI HTTP vs GridFTP : Multiple streams Not explicit in the HTTP protocol But needed for even higher performance – Especially in the WAN So we added it, with some semantics – Small wrapper around libcurl – PUT with ‘0 bytes’ && null content-range == end of write Submitted patch to libcurl to allow ssl session reuse among parallel requests

EMI INFSO-RI HTTP vs GridFTP: 3 rd Party Copies Implemented using WEBDAV COPY Requires proxy certificate delegation – Using gridsite delegation, with a small wrapper client Requires some common semantics to copy between SEs (to be agreed) – Common delegation portType location and port – No prefix in the URL ( just / )

EMI INFSO-RI DPM HTTP / DAV : Performance Xeon 4 Cores 2.27GHz 12 GB RAM 1 Gbit/s links Ongoing Evaluation No difference detected in LAN with different number of streams – But early results do show a big difference on the WAN lcg-cp configured to use gridftp File registration & transfer times considered in both cases

EMI INFSO-RI DPM HTTP / DAV : FTS Usage Example of FTS usage

EMI INFSO-RI DPM Beta: NFS 4.1 / pNFS Overview, Performance

EMI INFSO-RI NFS 4.1/pNFS: Why? Industry standard (IBM, NetApp, EMC, …) No vendor lock-in Free clients (with free caching) Strong security (GSSAPI) Parallel data access Easier maintenance … But you know all this by now…

EMI INFSO-RI NFS 4.1/pNFS: Overview CLIENT METADATA SERVER DISK SERVER(s) OPEN 1 LAYOUTGET 2 3 GETDEVICEINFO 4 OPEN 5 READ / WRITE 6 CLOSE 7

EMI INFSO-RI NFS4.1 / pNFS: Performance IOZONE Results Server – Xeon 4 Cores 2.27GHz – 12 GB RAM – 1 Gbit/s links Client – Dual core – 2 GB RAM – 100 Mbit/s link Ongoing Evaluation

EMI INFSO-RI NFS4.1 / pNFS: Performance NFS vs RFIO Ongoing Evaluation RFIO read misbehaving in this test… investigating Server – Xeon 4 Cores 2.27GHz – 12 GB RAM – 1 Gbit/s links Client – Dual core – 2 GB RAM – 100 Mbit/s link 8 KB block sizes

EMI INFSO-RI Conclusion fixes many scalability and performance issues – But we continue testing and improving Popular requests coming in next versions – Accounting, quotas, easier replication Beta components getting to production state – Standards compliant data access – Simplified setup, configuration, maintenance – Metadata consistency and synchronization And much more extensive testing – Performance test suites, regular large scale tests