CernVM-FS – Best Practice to Consolidate Global Software Distribution Catalin CONDURACHE, Ian COLLIER STFC RAL Tier-1 ISGC15, Taipei, March 2015.

Slides:



Advertisements
Similar presentations
The National Grid Service and OGSA-DAI Mike Mineter
Advertisements

Delivering Experiment Software to WLCG sites A new approach using the CernVM Filesystem (cvmfs) Ian Collier – RAL Tier 1 HEPSYSMAN.
Technology Steering Group January 31, 2007 Academic Affairs Technology Steering Group February 13, 2008.
15th January, NGS for e-Social Science Stephen Pickles Technical Director, NGS Workshop on Missing e-Infrastructure Manchester, 15 th January, 2007.
L. Arrabito, D. Bouvet, X. Canehan, P. Girard, Y. Perret, S. Poulat, R. Rumler SW distribution tests at Lyon Pierre Girard Luisa Arrabito, David Bouvet.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
Security Middleware and VOMS service status Andrew McNab Grid Security Research Fellow University of Manchester.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Introduction to CVMFS A way to distribute HEP software on cloud Tian Yan (IHEP Computing Center, BESIIICGEM Cloud Computing Summer School.
A short introduction to the Worldwide LHC Computing Grid Maarten Litmaath (CERN)
PanDA Multi-User Pilot Jobs Maxim Potekhin Brookhaven National Laboratory Open Science Grid WLCG GDB Meeting CERN March 11, 2009.
Configuration Management with Cobbler and Puppet Kashif Mohammad University of Oxford.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
Global Grid Forum GridWorld GGF15 Boston USA October Abhishek Singh Rana and Frank Wuerthwein UC San Diegowww.opensciencegrid.org The Open Science.
Predrag Buncic (CERN/PH-SFT) WP9 - Workshop Summary
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
GridPP Dirac Service The 4 th Dirac User Workshop May 2014 CERN Janusz Martyniak, Imperial College London.
Changes to CernVM-FS repository are staged on an “installation box" using a read/write file system interface. There is a dedicated installation box for.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
HEPiX IPv6 Working Group David Kelsey GDB, CERN 11 Jan 2012.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
Andrew McNabGrid in 2002, Manchester HEP, 7 Jan 2003Slide 1 Grid Work in 2002 Andrew McNab High Energy Physics University of Manchester.
Catalin Condurache STFC RAL Tier-1 GridPP OPS meeting, 10 March 2015.
The GridPP DIRAC project DIRAC for non-LHC communities.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES CVMFS deployment status Ian Collier – STFC Stefan Roiser – CERN.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Testing CernVM-FS scalability at RAL Tier1 Ian Collier RAL Tier1 Fabric Team WLCG GDB - September
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
CernVM-FS Infrastructure for EGI VOs Catalin Condurache - STFC RAL Tier1 EGI Webinar, 5 September 2013.
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
LCG Issues from GDB John Gordon, STFC WLCG MB meeting September 28 th 2010.
The GridPP DIRAC project DIRAC for non-LHC communities.
WLCG Operations Coordination report Maria Alandes, Andrea Sciabà IT-SDC On behalf of the WLCG Operations Coordination team GDB 9 th April 2014.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
ETICS An Environment for Distributed Software Development in Aerospace Applications SpaceTransfer09 Hannover Messe, April 2009.
SAM Status Update Piotr Nyczyk LCG Management Board CERN, 5 June 2007.
Andrea Manzi CERN EGI Conference on Challenges and Solutions for Big Data Processing on cloud 24/09/2014 Storage Management Overview 1 24/09/2014.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
STFC in INDIGO DataCloud WP3 INDIGO DataCloud Kickoff Meeting Bologna April 2015 Ian Collier
II EGEE conference Den Haag November, ROC-CIC status in Italy
The HEPiX IPv6 Working Group David Kelsey (STFC-RAL) EGI OMB 19 Dec 2013.
Considerations on Using CernVM-FS for Datasets Sharing Within Various Research Communities Catalin Condurache STFC RAL UK ISGC, Taipei, 18 March 2016.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Update on Service Availability Monitoring (SAM) Marian Babik, David Collados,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Accounting Update John Gordon. Outline Multicore CPU Accounting Developments Cloud Accounting Storage Accounting Miscellaneous.
CVMFS Alessandro De Salvo Outline  CVMFS architecture  CVMFS usage in the.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
CernVM-FS vs Dataset Sharing
Large Output and Shared File Systems
ATLAS Cloud Operations
HEPiX Spring 2014 Annecy-le Vieux May Martin Bly, STFC-RAL
StratusLab Final Periodic Review
StratusLab Final Periodic Review
Introduction to CVMFS A way to distribute HEP software on cloud
Cristina del Cano Novales STFC - RAL
IPv6 update Duncan Rand Imperial College London
Presentation transcript:

CernVM-FS – Best Practice to Consolidate Global Software Distribution Catalin CONDURACHE, Ian COLLIER STFC RAL Tier-1 ISGC15, Taipei, March 2015

Outline Introduction Brief history CernVM-FS deployment at RAL EGI CernVM-FS Task Force EGI CernVM-FS infrastructure CernVM-FS impact on user communities biomed and CernVM-FS ISGC15, Taipei, March 2015

Introduction – CernVM-FS ? Read-only, globally distributed filesystem, optimized to access experiment software and conditions data Built using standard technologies (fuse, sqlite, http, squid and caches) Files and directories are hosted on standard web servers and mounted in the universal /cvmfs namespace ISGC15, Taipei, March 2015

Introduction – CernVM-FS ? Software needs one single installation, then it is available at any site with CernVM-FS client installed and configured Repository meta-data and files are downloaded on demand and locally cached ISGC15, Taipei, March 2015

Brief History – CernVM-FS Last 5 years distribution of experiment software and conditions data to WLCG sites has been changed massively –CernVM-FS became the primary method –No need of local installation jobs local SW tags (!) (heavy loaded) file servers –It relies on a robust decentralised network of repositories replicas ISGC15, Taipei, March 2015

Brief History – CernVM-FS In parallel the use of CernVM-FS in communities outside WLCG has been increasing steadily –Growing number of repositories and CernVM-FS servers around the world Over 2 years ago RAL Tier-1 started a non-LHC Stratum-0 service ISGC15, Taipei, March 2015

Non-LHC CernVM-FS Deployment at RAL Supported by GridPP UK project Initially (Dec’12) for UK VOs, then extended to international small VOs (2+ NGIs) Local installation jobs used to automatically publish the Stratum-0 – now obsolete Replaced by web interface used by VO Software Grid Managers to upload, unpack tarballs and publish – now less used ISGC15, Taipei, March 2015

Non-LHC CernVM-FS Deployment at RAL Recently Grid Security Interface (GSI) added to transfer and process tarballs and publish –Faster and easier –Allows a programmatic way to transfer and process the tarballs –Initially based only on DN access, then VOMS Roles access also enabled – January 2015 –Software available at WN level after max 3 hours ISGC15, Taipei, March 2015

Non-LHC CernVM-FS Deployment at RAL At the moment 21 repos (~500 GB) are hosted and published at RAL Separate Stratum-1 service for non-LHC VOs –Initially shared the WLCG Stratum-1 –Recently deployed HA 2-node cluster – plans for a consolidated Stratum-1 at RAL (WLCG and non-WLCG) ISGC15, Taipei, March 2015

Software Installation Mechanism at RAL ISGC15, Taipei, March 2015 /home/augersgm /home/biomedsgm. /home/t2ksgm GSI Interface CVMFS Uploader Stratum-0 GSIssh/scp DN credentials VOMS Role credentials VO Software Grid Managers /cvmfs/auger.egi.eu /cvmfs/biomed.egi.eu. /cvmfs/t2k.egi.eu

EGI CernVM-FS Task Force EGI initiative to establish a CernVM-FS infrastructure that allows EGI Virtual Organizations to use it as a standard method of distribution of their software at grid sites Kick-off meeting – August 2013 Members: representatives from NGIs and other collaborations, VO SGMs, system admins at sites ISGC15, Taipei, March 2015

EGI CernVM-FS Task Force To promote the use of CernVM-FS technology amongst user communities To create a network of sites providing CernVM-FS services (Stratum-0, Stratum-1, Squid proxies) To encourage cooperation with other organizations (OSG, WLCG, regional grid infrastructures) by cross-replicating repositories for VOs supported by multiple collaborations ISGC15, Taipei, March 2015

EGI CernVM-FS Infrastructure The model follows the CernVM-FS topology for WLCG ISGC15, Taipei, March 2015

EGI CernVM-FS Infrastructure ‘relaxed’ topology Stratum-0 are disjoint and represent the source repositories where software is installed by VOs Stratum-0, Stratum-1 can be geographically co- located or not Stratum-1 can replicate an entire Stratum-0 or can only partially replicate – the ’relaxed’ model It partially makes use of the existent hierarchy of proxy servers used for LHC software distribution ISGC15, Taipei, March 2015

EGI CernVM-FS Infrastructure ISGC15, Taipei, March 2015 Stratum-0 NIKHEF nikhef.nl Stratum-1 NIKHEF Proxy Hierarchy Stratum-1 CERN Proxy Hierarchy Stratum-0 DESY desy.de Stratum-1 RAL Stratum-0 RAL egi.eu Proxy Hierarchy Stratum-1 DESY Proxy Hierarchy

EGI CernVM-FS Extended Infrastructure ISGC15, Taipei, March 2015 Stratum-0 NIKHEF nikhef.nl Stratum-1 NIKHEF Proxy Hierarchy Stratum-1 CERN Proxy Hierarchy Stratum-0 DESY desy.de Stratum-1 RAL Stratum-0 RAL egi.eu Proxy Hierarchy Stratum-1 DESY Proxy Hierarchy Stratum-1 ASGC Stratum-1 TRIUMF

EGI CernVM-FS Infrastructure 31 software repositories currently hosted –HEP and non-HEP –Stratum-0s at RAL, NIKHEF, DESY –5 repos at the time of kick-off –BIG change in two years! ISGC15, Taipei, March 2015

EGI CernVM-FS Infrastructure ‘egi.eu’ domain –Its importance recognized at the CernVM-FS WG meeting (March 2014) –Taken into account by latest CernVM-FS packages releases public key and CernVM-FS domain configuration now part of standard installation (as for ‘cern.ch’) – cvmfs-keys v1.5 no longer CERN-centric configuration – cvmfs v new puppet module available ISGC15, Taipei, March 2015

CernVM-FS Impact on User Communities Encouraged by the expansion of CernVM-FS infrastructure, more VOs have identified CernVM- FS as a facility that suited better their needs –HEP (auger, t2k.org, hyperk.org, glast.org) –Space, Natural and Life Sciences (biomed, km3net.org, enmr.eu, pheno) –EGI fedcloud community – exploratory discussions But software distribution to grid and cloud sites still an obstacle for some non-LHC VOs ISGC15, Taipei, March 2015

CernVM-FS Impact on User Communities Deployment issues –VOs are lacking manpower to progress more –Technology and its advantages are understood, just they do not have time –Subgroups within same VO are using subsets of software at specific sites coordination needed to consolidate software under single CernVM-FS space –Sometimes VOs wait too long for sites to help them with initial tests before rolling out more widely ISGC15, Taipei, March 2015

biomed and CernVM-FS biomed - large scale international and multi- disciplinary VO supporting communities from the Life Sciences sector Divided into three sectors: medical imaging, bioinformatics and drug discovery Member of the Life Science Grid Community –Other members: vlemed, lsgrid, vo.eu-decide.eu, enmr.eu Openly accessible to academics and to private companies for non-commercial purposes ISGC15, Taipei, March 2015

biomed and CernVM-FS biomed - Status at September 2013 –~80 supporting sites –~50 sites running significant production –3M jobs during Sep’13 Many biomed sub-groups installed various software packages at supporting sites –‘biomedsgm’ installation jobs –Need for NFS (or AFS) space at sites –Distribution of new SW releases was time consuming ISGC15, Taipei, March 2015

biomed and CernVM-FS Timeline –Sep’13 – initial discussions (EGI TF Madrid) –Oct’13 – Mar’14 Stratum-0 repository at RAL local install jobs used initially for upload web upload interface and GSI interface tested work to prove the concept within biomed communities first test jobs run at RAL using CernVM-FS space – Jan’14 –Apr’14 supporting sites asked (GGUS, personal contact) to enable biomed ISGC15, Taipei, March 2015

biomed and CernVM-FS Timeline –Jul’14 GSI as standard method to upload biomed shifters invited to use CernVM-FS –Sep’14 GGUS campaign for more sites to enable CernVM-FS support –Nov – Dec’14 44 out of 90 sites provide CernVM-FS VO-biomed-CVMFS tag used to identify supporting sites biomed community informed about CernVM-FS availability ISGC15, Taipei, March 2015

biomed and CernVM-FS ISGC15, Taipei, March 2015 Current status –75 sites running biomed jobs –90 out of 169 CEs advertise VO-biomed-CVMFS tag

Summary – CernVM-FS ISGC15, Taipei, March 2015 Success story Critical service for WLCG –soon to become critical for others Non-LHC Virtual Organizations are successfully following the model –biomed, enmr.eu, auger –increasing number of EGI VO repos –75 sites running biomed jobs –90 out of 169 CEs advertise VO-biomed-CVMFS tag

ISGC15, Taipei, March 2015 Thank you! (more) Questions?