EGEE-II INFSO-RI-031688 Enabling Grids for E-sciencE www.eu-egee.org EGEE and gLite are registered trademarks gLite Data Management Components Presenter.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel - WP2. WP2 Overview.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
DPM Name Server (DPNS) Namespace Authorization Location of physical files DPM Server Requests queuing and processing Space Management SRM Servers v1.1,
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
EGEE-II INFSO-RI Enabling Grids for E-sciencE gLite Data Management System Yaodong Cheng CC-IHEP, Chinese Academy.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
EGEE-III INFSO-RI Enabling Grids for E-sciencE The Medical Data Manager : the components Johan Montagnat, Romain Texier, Tristan.
Data Management The GSM-WG Perspective. Background SRM is the Storage Resource Manager A Control protocol for Mass Storage Systems Standard protocol:
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware Data Management in gLite.
INFSO-RI Enabling Grids for E-sciencE Distributed Metadata with the AMGA Metadata Catalog Nuno Santos, Birger Koblitz 20 June 2006.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
INFSO-RI Enabling Grids for E-sciencE Experiences with LFC and comparison with RNS Erwin Laure Jean-Philippe.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Management cluster summary Krzysztof Nienartowicz JRA1 All Hands meeting, Helsinki.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks MSG - A messaging system for efficient and.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Data management in LCG and EGEE David Smith.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 in DPM Sophie Lemaitre Jean-Philippe.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Regional Nagios Emir Imamagic /SRCE EGEE’09,
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
INFSO-RI Enabling Grids for E-sciencE University of Coimbra GSAF Grid Storage Access Framework Salvatore Scifo INFN of Catania EGEE.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Management cluster summary David Smith JRA1 All Hands meeting, Catania, 7 March.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data management in EGEE.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
INFSO-RI Enabling Grids for E-sciencE University of Coimbra gLite 1.4 Data Management System Salvatore Scifo, Riccardo Bruno Test.
INFSO-RI Enabling Grids for E-sciencE University of Coimbra Data Management System gLite – LCG – FiReMan Salvatore Scifo INFN Catania.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
EGEE-II INFSO-RI Enabling Grids for E-sciencE WLCG File Transfer Service Sophie Lemaitre – Gavin Mccance Joint EGEE and OSG Workshop.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
INFSO-RI Enabling Grids for E-sciencE FTS Administrators Tutorial for Tier-2s Paolo Badino
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
User Domain Storage Elements SURL  TURL LFC Domain (LCG File Catalogue) SA1 – Data Grid Interoperation Enabling Grids for E-sciencE EGEE-III INFSO-RI
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
2 nd EGEE/OSG Workshop Data Management in Production Grids 2 nd of series of EGEE/OSG workshops – 1 st on security at HPDC 2006 (Paris) Goal: open discussion.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
Data Management & Information Systems
EGEE Data Management Services
Jean-Philippe Baud, IT-GD, CERN November 2007
gLite Basic APIs Christos Filippidis
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
Status of the SRM 2.2 MoU extension
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
Introduction to Data Management in EGI
GFAL 2.0 Devresse Adrien CERN lcgutil team
Ákos Frohner EGEE'08 September 2008
Data Management cluster summary
Architecture of the gLite Data Management System
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks gLite Data Management Components Presenter : Markus W. Schulz, CERN, IT-GD On behalf of the Grid Data Management Team CHEP07, 2-7 Sept 2007, Victoria BC Canada

Enabling Grids for E-sciencE EGEE-II INFSO-RI CHEP07, 2-7 Sept 2007, Victoria BC Canada 2 WLCG & EGEE Large Hadron Collider (LHC) –15 PB/year to be stored at rates up to 1.5GB/sec (ALICE) and MB/sec (ATLAS, CMS, LHCb) –Data sharing : ~500 Institutes, 5000 physicists, computer scientists and engineers Need an infrastructure to support data analysis, sharing and storage Worldwide LHC Computing Grid (WLCG) –Data processing based on a Tier-Model (Tier-0, Tier-1,Tier-2) –Use of Open Science Grid (OSG), EGEE (EGEE), NDGF,+OSGEGEE), Will focus on EGEE data management contribution Enabling Grids for E-SciencE (EGEE) –Grid infrastructure for science (HEP, medicine, astronomy,...) –240+ sites, 45+ countries –Uses gLite as a lightweight, open source middleware distribution

Enabling Grids for E-sciencE EGEE-II INFSO-RI gLite architecture Service Oriented Architecture interoperability between grids support of grid standards flexible exploitation of the grid services according to specific needs CHEP07, 2-7 Sept 2007, Victoria BC Canada 3

Enabling Grids for E-sciencE EGEE-II INFSO-RI EGEE Data Management lcg_utils FTS Vendor Specific APIs GFAL CatalogingStorageData transfer Data Management User ToolsVO Frameworks (RLS) LFC SRM (Classic SE) gridftp RFIO Information System/Environment Variables

Enabling Grids for E-sciencE EGEE-II INFSO-RI LFC LCG File Catalog LHC Computing Grid File Catalog Large Hadron Collider Computing Grid File Catalog

Enabling Grids for E-sciencE EGEE-II INFSO-RI The LFC stores mappings between –Users’ file names and file locations on the Grid –Stores Permissions and –Ownership –Simple metadata Provides a hierarchical name space Supports GSI security model –Including VOMS based ACLs  Very fine grained control  Implementation based on virtual IDs –Soon: encrypted channels Simple DLI interface –Data Location Interface –GUID Location –Integration with WMS&RBs LCG “File” Catalog … File replica 2 GUID File replica 1 File replica m LFC file name 1 LFC file name n … “Replicas” are “Copies” All files are “Write Once” /grid /vo /data file

Enabling Grids for E-sciencE EGEE-II INFSO-RI MySQL and ORACLE back-ends –Ensures scalability and allows small scale deployment –Read only replication of catalogue (awaiting wider deployment) Multi-threaded C server –Supports multiple instances for load balancing Thread-safe C clients –Python & Perl bindings –Command line interface Supports sessions to avoid authentication costs –GSI is very expensive! Bulk methods to reduce the number of round trips – Under test by ATLAS --> 20 times faster Widely used in EGEE: –largest LFC instance contains 8 millions entries See poster session 2 “LFC status and recent developments” for more details LFC CHEP07, 2-7 Sept 2007, Victoria BC Canada 7

Enabling Grids for E-sciencE EGEE-II INFSO-RI DPM (Disk Pool Manager) DPM Disk Pool Manager SRM Storage Resource Manager

Enabling Grids for E-sciencE EGEE-II INFSO-RI Storage Element Storage Resource Manager (SRM) –Standard that hides the storage system implementation (disk or active tape) –handles authorization –Web service based on gsoap –translates SURLs (Storage URL) to TURLs (Transfer URLs) –disk-based: DPM, dCache,Storm; tape-based: Castor, dCache –SRM-2.2  Space tokens (manage space by VO/USER), advanced authorization,  Better handling of directories, lifetime, See talk “SRM 2.2 definition and testing experience” for more details (today) File I/O: posix-like access from local nodes or the grid  GFAL (Grid File Access Layer)

Enabling Grids for E-sciencE EGEE-II INFSO-RI /vo DPM: user's point of view /dpm /domain /home DPM head node file (uid, gid1, …) DPM disk servers DPM Name Server –Namespace, Authorization, file location Disk Servers –Physical files Direct data transfer from/to disk server (no bottleneck) External transfers via gridFTP Supports SRM –SRM-1.1, 2.1, 2.2 (since 1.6.3) CLI, C API, SRM-enabled client, etc. data transfer

Enabling Grids for E-sciencE EGEE-II INFSO-RI DPM Addresses the storage needs of Tier-2 and smaller sites Focus on easy setup and maintenance Multi-threaded C implementation Name server DB –Keeps track of the status of files and their physical locations –MySQL and ORACLE back ends  Simplifies integration in existing local DB infrastructure  Ensures scalability –Shares code with LFC --> fix once run twice! Thread-safe C client and command line interface –http/https DPM browser ( implemented, very soon to be released) –users and site managers interact with DPM at different levels GSI and VOMS based authorization and fine grained ACLs –Implemented via virtual IDs -> no excessive use of pool accounts –Pool access control on VO basis CHEP07, 2-7 Sept 2007, Victoria BC Canada 11

Enabling Grids for E-sciencE EGEE-II INFSO-RI DPM Different file access and transfer protocols Secure Remote File Input/Output (RFIO) –Secure file transfer and manipulation. –Implementation of thread-safe C client and a command line interface –Support of streaming mode. GSIFTP allows remote file transfer –New gridftp plugin is implemented to support gridFTP-2 Xrootd: useable but still limited –no support of grid/voms certificates yet https/http: web access based on Apache. –Protocol http or https can be specified at transfer time See poster session 2 “DPM Status and Next Steps” for more details CHEP07, 2-7 Sept 2007, Victoria BC Canada 12

Enabling Grids for E-sciencE EGEE-II INFSO-RI Virtual ID concept CHEP07, 2-7 Sept 2007, Victoria BC Canada 13

Enabling Grids for E-sciencE EGEE-II INFSO-RI LFC & DPM deployment status EGEE Catalog –110 LFCs in production  37 central LFCs  73 local LFCs EGEE Storage Elements –CASTOR –dCache –DPM  96 DPMs in production  Supporting 135 VOs LFC and DPM –Stable and reliable production quality services –Well established services –Require low support effort from administrators and developers Storage Element instances published in EGEE’s Top BDII Data volume distribution looks quite different

Enabling Grids for E-sciencE EGEE-II INFSO-RI Medical Data Management Problem : Medical institutes request data storage encryption Use of the DICOM standard for medical image handling Image retrieval and storage from/in DICOM servers : security issues Solution : Extension of the data management tools (under way) File encryption on the fly, local decryption Use of HYDRA for split key management Use of the LFC to register/retrieve system data –Replicas location, filesize, … Use of srmv2 to get the turls Use of I/O protocols, gridftp to load medical images Access control based on VOMS See poster session 2 “Medical Data Management” for more details CHEP07, 2-7 Sept 2007, Victoria BC Canada 15

Enabling Grids for E-sciencE EGEE-II INFSO-RI DPM (Disk Pool Manager) Lcg-util & GFAL Grid File Access Layer

Enabling Grids for E-sciencE EGEE-II INFSO-RI Lcg-util/gfal features Purpose: Create the illusion of POSIX I/O –Shield users from complexity –Interact with the information system, catalogue, SRMs  Can be used with/without information system/ catalogue LCG-util : –Command line and C-Api –Covers most common use cases  Replication, catalogue interaction etc, –high level tool box Gfal: –Posix like C API for file access –SRMv2.2 support  user space tokens for retention policy (custodial/replica) & access latency (online/nearline) See Poster session 2 “lcg-util and gfal” for more details CHEP07, 2-7 Sept 2007, Victoria BC Canada 17

Enabling Grids for E-sciencE EGEE-II INFSO-RI Lcg-util/gfal architecture CHEP07, 2-7 Sept 2007, Victoria BC Canada 18

Enabling Grids for E-sciencE EGEE-II INFSO-RI DPM (Disk Pool Manager) FTS File Transfer Service

Enabling Grids for E-sciencE EGEE-II INFSO-RI FTS overview gLite File Transfer Service is a reliable data movement service (batch for file transfers) –FTS performs bulk file transfers between multiple sites –Transfers are made between any SRM-compliant storage elements (both SRM 1.1 and 2.2 supported) –It is a multi-VO service, used to balance usage of site resources according to the SLAs agreed between a site and the VOs it supports –VOMS aware

Enabling Grids for E-sciencE EGEE-II INFSO-RI FTS Why is it needed ? –For the user, the service it provides is the reliable point to point movement of Storage URLs (SURLs) and ensures you get your share of the sites’ resources –For the site manager, it provides a reliable and manageable way of serving file movement requests from their VOs and an easy way to discover problems with the overall service delivered to the users –For the VO production manager, it provides ability to control requests coming from his users Re-ordering, prioritization,… –The focus is on the “service” delivered to the user It makes it easy to do these things well with minimal manpower

Enabling Grids for E-sciencE EGEE-II INFSO-RI FTS: key points Reliability –It handles the retries in case of storage / network failures –VO customizable retry logic –Service designed for high-availability deployment Security –All data is transferred securely using delegated credentials with SRM / gridFTP –Service audits all user / admin operations Service and performance –Service stability: it is designed to efficiently use the available storage and network resources without overloading them –Service recovery: integration of monitoring to detect service-level degradation

Enabling Grids for E-sciencE EGEE-II INFSO-RI Service scale Designed to scale up to the transfer needs of very data intensive applications Currently deployed in production at CERN –Running the production WLCG tier-0 data export –Target rate is ~1 Gbyte/sec 24/7 –Over 9 petabytes transferred in last 6 months > 10 million files Also deployed at ~10 tier-1 sites running a mesh of transfers across WLCG –Inter-tier1 and tier-1 to tier-2 transfers –Each tier-1 has transferred around 0.2 – 0.5 petabytes of data

Enabling Grids for E-sciencE EGEE-II INFSO-RI FTS New features in FTS 2.0 –Better security model (certificate delegation) –Support for SRM v2 –More administrative tools, more advanced monitoring features to make it easier to operate the overall service –Soon:  Better support for clouds and channel sets  Black/White-listing SEs  Better integration with VO workflow management Call backs, hooks in the state machine Focus continues upon service monitoring and easing the service operations together with closer integration of FTS with experiment software frameworks see talk “Building the WLCG file transfer service” for more details CHEP07, 2-7 Sept 2007, Victoria BC Canada 24

Enabling Grids for E-sciencE EGEE-II INFSO-RI Is this all? Definitely NO!!! The AMGA meta data catalogue ( by Birger Koblitz) –Widely used by experiments –Is in the process to be integrated in the gLite distribution Many data management tools and services developed by Vos Lessons learned –A DM stack can only be developed with production feedback –The right balance between exposing details and hiding is hard to find –There will be more to do CHEP07, 2-7 Sept 2007, Victoria BC Canada 25

Enabling Grids for E-sciencE EGEE-II INFSO-RI Summary Current status Data Management framework is usable LFC, FTS, DPM and lcg-util/gfal are used in production on a large scale Outlook & Future ACL synchronization between LFC and SEs Improvements to lcg-util/gfal –e.g. flexibility to work independently of the LFC Better tools to check consistency in DPM Extension of Xrootd to support grid/voms certificates Finish medical data management implementation DPM : quota on pools and accounting Operational improvements to the FTS Continue the dialog with the user communities to focus effort CHEP07, 2-7 Sept 2007, Victoria BC Canada 26

Enabling Grids for E-sciencE EGEE-II INFSO-RI LFC bulk methods //deleting replicas (if force=1) and LFNs if mapped to one of guid in the list of guids int DLL_DECL lfc_delfilesbyguid(int nbguids, const char **guids, int force, int *nbstatuses, int **statuses) //deleting replicas (if force=1) and LFNs (specified by the **paths) int DLL_DECL lfc_delfilesbyname(int nbfiles, const char **paths, int force, int *nbstatuses, int **statuses) //deleting replicas (if force=1) and LFNs if LFN starts with concat(path,pattern) int DLL_DECL lfc_delfilesbypattern(const char *path, const char *pattern, int force, int *nbstatuses, struct Cns_filestatus **statuses) //deleting replicas stored on the given SE and associated with one guid provided by the list of guids int DLL_DECL lfc_delreplicas(int nbguids, const char **guids, char *se, int *nbstatuses, int **statuses) //list the content of the directory + replica info if matches the pattern struct lfc_direnrep DLL_DECL * lfc_readdirxp(Cns_DIR *dirp, char *pattern, char *se) CHEP07, 2-7 Sept 2007, Victoria BC Canada 27