CVMFS Alessandro De Salvo 18-5-2011 Outline  CVMFS architecture  CVMFS usage in the.

Slides:



Advertisements
Similar presentations
Welcome to Middleware Joseph Amrithraj
Advertisements

Delivering Experiment Software to WLCG sites A new approach using the CernVM Filesystem (cvmfs) Ian Collier – RAL Tier 1 HEPSYSMAN.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
1 Worker Nodes Web Proxies Grid Site Repository Mirrors (Stratum 1) HTTP CernVM-FS fuse module on WNs recommended deployment Worker Nodes Web Proxies Grid.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
CVMFS AT TIER2S Sarah Williams Indiana University.
1 Oracle 9i AS Availability and Scalability Margaret H. Mei Senior Product Manager, ST.
1 The Google File System Reporter: You-Wei Zhang.
Introduction to CVMFS A way to distribute HEP software on cloud Tian Yan (IHEP Computing Center, BESIIICGEM Cloud Computing Summer School.
Interactive Job Monitor: CafMon kill CafMon tail CafMon dir CafMon log CafMon top CafMon ps LcgCAF: CDF submission portal to LCG resources Francesco Delli.
Daniela Anzellotti Alessandro De Salvo Barbara Martelli Lorenzo Rinaldi.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
ROOT and Federated Data Stores What Features We Would Like Fons Rademakers CERN CC-IN2P3, Nov, 2011, Lyon, France.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Usage of virtualization in gLite certification Andreas Unterkircher.
Changes to CernVM-FS repository are staged on an “installation box" using a read/write file system interface. There is a dedicated installation box for.
1 MSRBot Web Crawler Dennis Fetterly Microsoft Research Silicon Valley Lab © Microsoft Corporation.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Tools and techniques for managing virtual machine images Andreas.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Using CVMFS to serve site software Sarah Williams Indiana University 2/01/121.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES CVMFS deployment status Ian Collier – STFC Stefan Roiser – CERN.
Testing CernVM-FS scalability at RAL Tier1 Ian Collier RAL Tier1 Fabric Team WLCG GDB - September
2nd ASPERA Workshop May 2011, Barcelona, Spain P. Mato /CERN.
CernVM-FS Infrastructure for EGI VOs Catalin Condurache - STFC RAL Tier1 EGI Webinar, 5 September 2013.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Status of CVMFS for ALICE Deployment timeline ALICE © | Offline Week | June 2013| Predrag Buncic.
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
Replicazione e QoS nella gestione di database grid-oriented Barbara Martelli INFN - CNAF.
CernVM-FS – Best Practice to Consolidate Global Software Distribution Catalin CONDURACHE, Ian COLLIER STFC RAL Tier-1 ISGC15, Taipei, March 2015.
Advancing CernVM-FS and its Development Infrastructure José Molina Colmenero CERN EP-SFT.
Considerations on Using CernVM-FS for Datasets Sharing Within Various Research Communities Catalin Condurache STFC RAL UK ISGC, Taipei, 18 March 2016.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Predrag Buncic, CERN/PH-SFT The Future of CernVM.
Parrot and ATLAS Connect
Jean-Philippe Baud, IT-GD, CERN November 2007
CernVM-FS vs Dataset Sharing
Sanjay Ghemawat, Howard Gobioff, Shun-Tak Leung
Dynamic Extension of the INFN Tier-1 on external resources
The EDG Testbed Deployment Details
StoRM: current status and developments
High Availability 24 hours a day, 7 days a week, 365 days a year…
System Center 2012 Configuration Manager
Database Replication and Monitoring
Virtualisation for NA49/NA61
Dag Toppe Larsen UiB/CERN CERN,
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Dag Toppe Larsen UiB/CERN CERN,
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
Future of WAN Access in ATLAS
Service Challenge 3 CERN
Database Readiness Workshop Intro & Goals
Introduction to CVMFS A way to distribute HEP software on cloud
Virtualisation for NA49/NA61
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Generator Services planning meeting
ATLAS Software Installation redundancy Alessandro De Salvo Alessandro
STORM & GPFS on Tier-2 Milan
The ATLAS software in the Grid Alessandro De Salvo <Alessandro
Quattor Usage at Nikhef
Torrent-based software distribution
Implementing CVMFS server using an union file system
ETICS Services Management
Virtualization in the gLite Grid Middleware software process
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
Chapter 15: File System Internals
Presentation transcript:

CVMFS Alessandro De Salvo Outline  CVMFS architecture  CVMFS usage in the LHC experiments  Performance tests A. De Salvo – 18 Dec 2011

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba CVMFS basics ATLAS and LHCb moving to a dynamic software distribution model via CVMFS (CernVMFileSystem) Virtual software installation by means of an HTTP File System Distribution of read-only binaries Files and file meta data are downloaded on demand and locally cached Self-contained (e. g. /cvmfs/atlas.cern.ch), does not interfere with the base system

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Content addressable storage Data Store Compressed Chunks (Files) Eliminates Duplicates File Catalog Directory Structure Symlinks SHA1 of Regular Files Digitally Signed Time to Live Nested Catalogs Immutable Files, trivial to check for corruption

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba CVMFS Clients Mounted in the remote nodes via the fuse module Local cache for faster access Benefits of a squid hierarchy to guarantee performance, scalability and reliability Same squid type as the one used for Frontier

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba CVMFS Backends Mirror servers Web servers listening on port 80, 8000 Proxy servers Local load-balanced Squid forward proxy (SL5 Squid)

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Fuse Module Namespace: /cvmfs/ e. g. /cvmfs/atlas.cern.ch One process per repository + watchdog One mount per repository Exclusive cache directory per repository Mount scripts Setup environment (number of file descriptors, access rights,... ) Used by autofs on /cvmfs Used by system mount mount -t cvmfs atlas.cern.ch /cvmfs/atlas.cern.ch Diagnostics service cvmfs probe cvmfs_config chksetup cvmfs_config showconfig cvmfs_fsck cvmfs-talk, connect to running instance CVMFS client tools

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Proxies SL5 Squid, Load-Balancing + Fail-Over e. g. CVMFS_HTTP_PROXY="A|B|C" as of next revision: semantics like Frontier/Squid Mirrors Fail-Over, set order manually in /etc/cvmfs/domain.d/cern.ch.local Mirrors at CERN, RAL, BNL operational by the end of the month CVMFS Web Caches

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Proxies SL5 Squid, Load-Balancing + Fail-Over e. g. CVMFS_HTTP_PROXY="A|B|C" as of next revision: semantics like Frontier/Squid Mirrors Fail-Over, set order manually in /etc/cvmfs/domain.d/cern.ch.local Mirrors at CERN, RAL, BNL operational by the end of the month CVMFS Web Caches

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Release manager machine acts as "template". Maintains shadow tree and CernVM-FS repository Publishing ① Install and test new software ② Synchronize changes with repository ③ Sign file catalogs ④ Create new file system snapshot CVMFS Backend storage Steps 2–4: Taken care of by CernVM-FS server tools

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Known users: CernVM, ATLAS U.S. Tier3s, RAL, PIC, NIKHEF, CERN, Wuppertal, QMUL, Munich, Lancester, Dortmund, JINR,... /cvmfs/atlas.cern.ch “Production” Software 5 release managers, 24 releases SLC releases SLC5 590 GB, 11 Million files, 16 Million entries (shadow) 85 GB and 1.5 Million files (repository) New production server with the final setup almost ready Populated, but not yet tested Will expose all the software structure as in current grid sites /cvmfs/atlas-condb.cern.ch — ATLAS Condition Flat Files Release manager machine hosted by CERN IT Automatic update several times a day 30 GB, files, directories, symlinks (shadow tree) 30 GB, files (repository) Only fraction of all conditions data /cvmfs/atlas-nightlies.cern.ch — ATLAS Nightlies Work in progress Integrated with the current Installation System CVMFS sites are used by the installation system transparently, aside of sites using a different FS ATLAS on CVMFS

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba CVMFS test at CNAF (LHCb)

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Tests performed Data Intensive Reading ~40 GB of software libraries (all the LHCb repository) from ~200 Worker Nodes concurrently find /opt/lhcb/lib/lhcb –name "*.so" -type f -exec cat {} >/dev/null \; Metadata intensive Initialization (SetupProject) and startup of the LHCb analysis program from ~200 Worker Nodes concurrently Very similar to the procedure used by ATLAS Currently implying a huge number of metadata lookup CVMFS tests at CNAF (LHCb): test definition

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Data intensive setup With squid caches unloaded As soon as the files are requested they are downloaded synchronously from CERN then served to the Worker Nodes Typical access mode of an access to a new software release With squid caches loaded The files are already present in the disks of the squid servers Measuring the performance of the software distribution from the squid caches to the CVMFS clients CVMFS tests at CNAF (LHCb): test definition [2]

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Peak at 240 MB/s (half bandwidth) CVMFS tests at CNAF (LHCb): data-intensive test results Peak at 440 MB/s (full bandwidth)

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba Metadata intensive setup With squid caches and local disk caches unloaded With squid caches loaded and local disk caches unloaded With squid caches and local disk caches loaded Same test executed for reference on cNFS over GPFS Without memory caching (page cache) in the WNs With memory caching CVMFS tests at CNAF (LHCb): test definition [3]

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba CVMFS: good results in general, 10 -> 100 seconds on a heavy-duty case CVMFS tests at CNAF (LHCb): metadata-intensive test results cNFS has similar performance to CVMFS But different architecture and features

A. De Salvo – CVMFS – CCR Workshop, Isola d’Elba New CVMFS stable version (2.0) ready by the end of April ATLAS and LHCb are moving to CVMFS as main distribution method Better management More efficiency Good performance Already integrated with the current installation system (ATLAS) Useful links Download Yum Mailing list News Bug Tracker Credits Thanks to V. Vagnoni and the CNAF folks for the LHCb tests Conclusions