EGEE-II INFSO-RI-031688 Enabling Grids for E-sciencE www.eu-egee.org Data Management cluster summary Krzysztof Nienartowicz JRA1 All Hands meeting, Helsinki.

Slides:



Advertisements
Similar presentations
DPM Name Server (DPNS) Namespace Authorization Location of physical files DPM Server Requests queuing and processing Space Management SRM Servers v1.1,
Advertisements

EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 experience Sophie Lemaitre WLCG Workshop.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
EGEE-III INFSO-RI Enabling Grids for E-sciencE The Medical Data Manager : the components Johan Montagnat, Romain Texier, Tristan.
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
New DM Clients and retirement plans for old ones Alejandro Alvarez Ayllon on behalf of the DM Clients developers IT/SDC 11/12/2013.
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
Author - Title- Date - n° 1 Partner Logo WP5 Summary Paris John Gordon WP5 6th March 2002.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks JRA1 summary Claudio Grandi EGEE-II JRA1.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
INFSO-RI Enabling Grids for E-sciencE Scenarios for Integrating Data and Job Scheduling Peter Kunszt On behalf of the JRA1-DM Cluster,
CERN SRM Development Benjamin Coutourier Shaun de Witt CHEP06 - Mumbai.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Medical Data Manager 1 Dicom retrieval : overview of the DPM One command line to retrieve a file:
INFSO-RI Enabling Grids for E-sciencE gLite Middleware Status Frédéric Hemmer, JRA1 Manager, CERN On behalf of JRA1.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
The new FTS – proposal FTS status. EMI INFSO-RI /05/ FTS /05/ /05/ Bugs fixed – Support an SE publishing more than.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Data management in LCG and EGEE David Smith.
EMI INFSO-RI European Middleware Initiative (EMI) Alberto Di Meglio (CERN)
INFSO-RI Enabling Grids for E-sciencE /10/20054th EGEE Conference - Pisa1 gLite Configuration and Deployment Models JRA1 Integration.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks gLite Data Management Components Presenter.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Storage Accounting for Grid Environments Fabio Scibilia INFN - Catania.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks GSI with OpenSSL Vincenzo Ciaschini EGEE-3.
DMLite GridFTP frontend Andrey Kiryanov IT/SDC 13/12/2013.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Patch Preparation SA3 All Hands Meeting.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Management cluster summary David Smith JRA1 All Hands meeting, Catania, 7 March.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
The GridPP DIRAC project DIRAC for non-LHC communities.
Enabling Grids for E-sciencE EGEE-III-INFSO-RI EGEE and gLite are registered trademarks Francesco Giacomini JRA1 Activity Leader.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Architecture of LHC File Catalog Valeria Ardizzone INFN Catania – EGEE-II NA3/NA4.
SRM 2.2: experiment requirements, status and deployment plans 6 th March 2007 Flavia Donno, INFN and IT/GD, CERN.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
EGEE-II INFSO-RI Enabling Grids for E-sciencE WLCG File Transfer Service Sophie Lemaitre – Gavin Mccance Joint EGEE and OSG Workshop.
Enabling Grids for E-sciencE EGEE-II INFSO-RI Status of SRB/SRM interface development Fu-Ming Tsai Academia Sinica Grid Computing.
INFSO-RI Enabling Grids for E-sciencE FTS Administrators Tutorial for Tier-2s Paolo Badino
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
2 nd EGEE/OSG Workshop Data Management in Production Grids 2 nd of series of EGEE/OSG workshops – 1 st on security at HPDC 2006 (Paris) Goal: open discussion.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
INFSO-RI Enabling Grids for E-sciencE Security needs in the Medical Data Manager EGEE MWSG, March 7-8 th, 2006 Ákos Frohner on behalf.
EGEE Data Management Services
Jean-Philippe Baud, IT-GD, CERN November 2007
Status of the SRM 2.2 MoU extension
Gfal/lcg-util -> Gfal2/gfal2-util
Encrypted Data Store, Hydra & Delegation Interface
Data Management cluster summary
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Management cluster summary Krzysztof Nienartowicz JRA1 All Hands meeting, Helsinki 18 June 2007

Enabling Grids for E-sciencE EGEE-II INFSO-RI Outlook Disk Pool Manager status LHC File Catalog status File Transfer Service status Grid File Access Library and lcg_util status Biomed Glite restructuring Data Management Summary 2

Enabling Grids for E-sciencE EGEE-II INFSO-RI GSSD June 2007: DPM status and plans DPM current status DPM running in production at more than 120 sites –Austria, Belgium, Canada, France, Greece, Hungary, India, Italy, Japan, Holland, Poland, Russia, Spain, Switzerland, Taiwan, UK DPM serving 148 Virtual Organizations –HEP experiments, biomed, esr, fusion … Collaboration with NAREGI on interoperability Collaboration with Globus

Enabling Grids for E-sciencE EGEE-II INFSO-RI GSSD June 2007: DPM status and plans DPM I Functionality offered in current production release (1.6.4): –Management of disk space on geographically distributed disk servers –Management of name space (including ACLs) –Control interfaces:  socket, SRM v1.0, SRM v2.1, SRM v2.2 (no srmCopy) –Data access protocols:  secure RFIO, GsiFTP (Globus 2.4)

Enabling Grids for E-sciencE EGEE-II INFSO-RI GSSD June 2007: DPM status and plans DPM II Functionality offered in latest release (1.6.5) –Full support for permanent and volatile spaces –ACLs on disk pools –Improved version of dpm-qryconf –Recursive srmLs and srmRmdir –SL4 32 bits release including gridFTP 2 plugin –SL4 on 64 bits architecture (being tested at a few sites) –Status: certified – moving onto PPS today (17 June)

Enabling Grids for E-sciencE EGEE-II INFSO-RI GSSD June 2007: DPM status and plans DPM III SRM v2.2 interface –Basic tests and use case tests are ok –Current stress tests show that the system is stable –Being deployed at selected sites in UK and France –srmCopy and srmChangeSpaceForFiles still missing Xrootd plugin –Requested by ALICE –Prototype delivered in November 2006 –Currently being tested by ALICE –Could also be used by ATLAS and CMS later (ongoing discussions)

Enabling Grids for E-sciencE EGEE-II INFSO-RI LFC LFC in production at over 70 sites for 158 VOs Current LFC production version 1.6.4: –Support for secondary groups –Bulk queries LFC planning –SSL enabled CSec version exists –SSL session re-use via Csec or openssl, Csec lib prepared –More bulk operations Data Management Summary 7

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 8 FTS current status Current FTS production status –CERN and all WLCG T1 sites currently FTS v1.5 –> 10 petabytes exported from CERN since SC4 FTS 2.0 has been well tested on pilot service –> 500 terabytes transferred –All VOs have agreed to upgrade to FTS 2.0  Validation took a while but it was vital to ensure the continuity of the production service –Scheduled intervention this week to upgrade CERN-PROD –Release will be made available to T1 sites after 3-4 weeks

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 9 FTS II FTS 2.0 new features –Delegation of proxy from the client to the FTS service –Improved monitoring capabilities  Critical to the ‘overall transfer service’ operational stability  Much more data retained in the database, some new methods to access them in the admin API –Beta SRM 2.2 support  This is now being tested on the PPS –Better administration tools  Make it easier to run the service –Better database model  Improve the performance and scalability –Placeholders for future functionality  Minimise the impact of future upgrade interventions

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 10 FTS III FTS planning –Evolve the SRM 2.2 code as we understand the SRM 2.2 implementations (based on feedback from PPS) –Incrementally improve service monitoring  FTS will have the capacity to give very detailed measurements about the current service level and problems currently being observed with sites  Integration with experiment and operations dashboards  Design work ongoing –Site grouping in channel definition (“clouds”)  To make it easier to implement the computing models of CMS and ALICE  Code exists: to be tested on pilot service –Incrementally improve service administration tools –SRM/gridFTP split –Notification of job state changes Not planned –Not planning to produce a non-Oracle version  Sites with no Oracle license can use CPU/memory restricted Oracle XE –Not planning to produce a more complex data movement orchestration service (Data Scheduler)  No real driving requirements for this

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 11 GFAL and lcg_util A number of bug fixes and improvements Fixes on the way… –Improved error messages –Improved python interface –Improvements to SRM 2.2 functionality: list support, bulk requests (srmRm, etc) –VO argument now not needed if not necessary –VO name max length is now larger Planning –Shorter term: more requested fixes, cleaner SRM GFAL functions –Longer term: thread safe version of GFAL / lcg_util

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 12 BIOMED (background) BIOMED –BIOMED VO would like certain data to be ‘visible on the grid’, with the ability to replicate it to ordinary sites with ordinary SEs –Have their own storage and wire protocol called DICOM –It is required that data be encrypted (and anonymised) after leaving the DICOM server and before going elsewhere on the site or the grid. Currently BIOMED are –Using a modified version of dCache with gLite I/O and Fireman to provide access the DICOM service (all together called the DICOM SE) –Hydra is used as a keystore for encryption keys, with the hydra and EDS clients (which in turn use the gLite I/O client) to access the data, retrieve the key and decrypt or encrypt the data.

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 13 BIOMED requirements Requirement summary –Files/pictures are in the DICOM server –LFN can be study/series/ID, which can be used with DICOM (no need for hierarchical directories and custom file names) –SRM interface –Efficient access, e.g. gridFTP or https - either are OK –Mandatory encryption –ACL on each file or picture (synchronized between SEs) –Use cases are still not fully understood  Need some more iteration with NA4

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 14 DICOM integration with DPM Initial design complete, prototyping underway –Need to develop the DPM staging backend  Design understood, work pending –Need to write the DICOM staging backend plug-in, with a callout to Hydra, to copy between DICOM server and the DPM  Designed, prototyping underway Hydra client using GFAL rather than gLite I/O prepared

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 15 ACL synchronization –To synchronize ACL updates on replicas at arbitrary sites –We believe we will need to provide explicit synchronization between LFC, Hydra, (DICOM SE) and other SEs. –Design and architecture discussions. Work about to start.  Likely based on queuing toolsets (JMS)

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 16 gLite restructuring Rationalise gSoap dependencies –Currently upgrading all to 2.7.6b Remove unnecessary CGSI_gSOAP usage –Replace by openssl and GridSite –Done for all components it makes sense for Migrate to libxml2 –Work started Test DM code with VDT 1.6 –DPM/LFC done, lcg-utils + FTS to do General cleanup of build dependencies –Much work done.. ongoing. Significant effort spent on ETICS integration –ETICS build is now working for us 32 bit / 64 bit SLC4 builds now ~completed –32 bit SLC4 build OK. Final fixes for 64 bit build being done. –LFC/DPM tested on 64 bit. Other components to test.

Enabling Grids for E-sciencE EGEE-II INFSO-RI Data Management Summary 17 Summary Summary and plans –Build work completed: gLite restructuring underway –Various improvements  FTS 2.0 released  New functionality for LFC/DPM  Lcg_util / GFAL bug fixes and new methods  SRM 2.2 introduced and now in active testing –BIOMED design work done, prototyping underway As always… –Significant effort provided to support operational services, user support and bug fixes