Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław

Slides:



Advertisements
Similar presentations
EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
Advertisements

Xrootd and clouds Doug Benjamin Duke University. Introduction Cloud computing is here to stay – likely more than just Hype (Gartner Research Hype Cycle.
INFSO-RI An On-Demand Dynamic Virtualization Manager Øyvind Valen-Sendstad CERN – IT/GD, ETICS Virtual Node bootstrapper.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Tier-1 experience with provisioning virtualised worker nodes on demand Andrew Lahiff, Ian Collier STFC Rutherford Appleton Laboratory, Harwell Oxford,
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Installation and Integration of Virtual Clusters onto Pragma Grid NAIST Nara, Japan Kevin Lam 06/28/13.
1 port BOSS on Wenjing Wu (IHEP-CC)
AliEn uses bbFTP for the file transfers. Every FTD runs a server, and all the others FTD can connect and authenticate to it using certificates. bbFTP implements.
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Data production using CernVM and lxCloud Dag Toppe Larsen Belgrade
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
Nadia LAJILI User Interface User Interface 4 Février 2002.
Data production using CernVM and LxCloud Dag Toppe Larsen Warsaw,
Event Data History David Adams BNL Atlas Software Week December 2001.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
Quattor-for-Castor Jan van Eldik Sept 7, Outline Overview of CERN –Central bits CDB template structure SWREP –Local bits Updating profiles.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Usage of virtualization in gLite certification Andreas Unterkircher.
Infrastructure for QA and automatic trending F. Bellini, M. Germain ALICE Offline Week, 19 th November 2014.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
WLCG Overview Board, September 3 rd 2010 P. Mato, P.Buncic Use of multi-core and virtualization technologies.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
Maite Barroso - 10/05/01 - n° 1 WP4 PM9 Deliverable Presentation: Interim Installation System Configuration Management Prototype
Workload management, virtualisation, clouds & multicore Andrew Lahiff.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen Budapest
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
NA61 Collaboration Meeting CERN, December Predrag Buncic, Mihajlo Mudrinic CERN/PH-SFT Enabling long term data preservation.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
WP5 – Infrastructure Operations Test and Production Infrastructures StratusLab kick-off meeting June 2010, Orsay, France GRNET.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Virtualisation: status and plans Dag Toppe Larsen
Predrag Buncic, CERN/PH-SFT The Future of CernVM.
CHEP 2010 Taipei, 19 October Predrag Buncic Jakob Blomer, Carlos Aguado Sanchez, Pere Mato, Artem Harutyunyan CERN/PH-SFT.
HEPiX Virtualisation working group Andrea Chierici INFN-CNAF Workshop CCR 2010.
Virtual machines ALICE 2 Experience and use cases Services at CERN Worker nodes at sites – CNAF – GSI Site services (VoBoxes)
CERN IT Department CH-1211 Genève 23 Switzerland The CERN internal Cloud Sebastien Goasguen, Belmiro Rodrigues Moreira, Ewan Roche, Ulrich.
CERN Openlab Openlab II virtualization developments Havard Bjerke.
CernVM-FS vs Dataset Sharing
Cloud Technology and the NGS Steve Thorn Edinburgh University (Matteo Turilli, Oxford University)‏ Presented by David Fergusson.
Installation of the ALICE Software
Use of HLT farm and Clouds in ALICE
Volunteer Computing for Science Gateways
Virtualization and Clouds ATLAS position
Virtualisation for NA49/NA61
NA61/NA49 virtualisation:
Blueprint of Persistent Infrastructure as a Service
Dag Toppe Larsen UiB/CERN CERN,
Dag Toppe Larsen UiB/CERN CERN,
ATLAS Cloud Operations
WLCG experiments FedCloud through VAC/VCycle in the EGI
Introduction to CVMFS A way to distribute HEP software on cloud
Virtualisation for NA49/NA61
The ATLAS software in the Grid Alessandro De Salvo <Alessandro
Artem Trunov and EKP team EPK – Uni Karlsruhe
ETICS Services Management
Virtualization in the gLite Grid Middleware software process
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
Discussions on group meeting
WLCG Collaboration Workshop;
Cloud Computing R&D Proposal
Haiyan Meng and Douglas Thain
Module 01 ETICS Overview ETICS Online Tutorials
Production Manager Tools (New Architecture)
Presentation transcript:

Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław 05.03.2012

Outline Quick reminder of CERNVM Tasks Roadmap Input needed Each task in detail Roadmap Input needed 05.03.2012 NA61/NA49 meeting, Wrocław

CERNVM CERNVM is a Linux-distribution Addition software Designed specifically for virtual machines (VMs) Based on SLC (currently SLC5) Compressed image size ~300MB Both 32-bit and 64-bit versions Addition software “Standard” software via Conary package manager Experiment software via CVMFS Contextualisation: images adapted to experiment requirements during boot Data preservation: all images are permanently preserved 05.03.2012 NA61/NA49 meeting, Wrocław

CVMFS Distributed read-only file system for CERNVM (i.e. the same as AFS for LXPLUS) Can also be used by “real” machines (e.g. LXPLUS, grid) Files compressed and distributed via HTTP Global availability Central server, site replication via standard HTTP proxies Files decompressed and cached on (CERNVM) computer Can run without Internet access if all needed files are cached Mainly for experimental software, but also other “static” data (e.g. calibration data) Each experiment has a repository to store all versions of software Common software (e.g. ROOT) available from SFT repository 05.03.2012 NA61/NA49 meeting, Wrocław

Data preservation As technology evolves, no longer possible to run legacy software on modern platforms Must be preserved and accessible: Experiment data Experiment software Operating environment (operating system, libraries, compilers, hardware) Just preserving data and software is not enough Virtualisation may preserve operating environment 05.03.2012 NA61/NA49 meeting, Wrocław

CERNVM data preservation “Solution”: Experiment data stored on Castor Experiment software versions stored on CVMFS HTTP “lasting” technology Operation environments stored as CERNVM image versions Thus, a legacy version of CERNVM can be started as a VM, running a legacy version of experiment software Forward-looking approach (we start preserving now) 05.03.2012 NA61/NA49 meeting, Wrocław

Tasks Make experiment software available Facilitate batch processing Validate outputs On-demand virtual clusters Production reconstruction Reference cloud cluster Data bookkeeping web interface 05.03.2012 NA61/NA49 meeting, Wrocław

Make experiment software available NA61/NA49 software must be available on CVMFS for CERNVM to process data NA61 Legacy software chain installed Changes to be fed back to SVN SHINE Preparing to install Use ROOT from SFT repository Conary package manager to install other dependencies Have to create package for XZ, currently not available Will there be 64-bit version of SHINE, or will it always be 32-bit? Installation expected to be easier than for legacy chain Not “critical” until ready, but good to gain experience, and be prepared NA49 SLC4 development machine and repository set up Need expert support with actual installation 05.03.2012 NA61/NA49 meeting, Wrocław

Facilitate batch processing LXPLUS uses PBS batch system, CERNVM uses Condor New scripts prepared “Philosophical” differences PBS has separate script for each job Condor has common job description file Installation of legacy NA61 reconstruction chain recently completed Issues discovered, which requires modifications to scripts But no big issues 05.03.2012 NA61/NA49 meeting, Wrocław

Validate outputs Data processed on CERNVM/CVMFS have to produce same results as from LXPLUS/AFS A larger data set should be used for this testing As part of processing the data on CERNVM, one can automatically run ds_diff on the newly reconstructed data, and LXPLUS data copied from Castor “Easy” to add to Condor script Output from ds_diff must be checked by hand Make sure same versions of reconstruction software is used 05.03.2012 NA61/NA49 meeting, Wrocław

On-demand virtual clusters On boot, the VMs are set up (contextualised) with the configurations and software needed by the relevant experiment Environment (variables, etc.) Version of experimental software Version of OS image Hardware configuration (e.g. RAM) VMs can be discarded after the data is processed A script will create a virtual cluster with head node and a suitable number of worker nodes Cluster discarded when jobs are finished Initially command-line script Later controlled by data bookkeeping web interface 05.03.2012 NA61/NA49 meeting, Wrocław

Production reconstruction After outputs are validated, production reconstruction next step Cluster of “decent” size needed Need to submit ~50 VMs to process a large data set Reference cloud too small Need to negotiate with IT to use LXCLOUD (not- yet-public CERN cloud) CERN already has a large number of internal virtual machines 05.03.2012 NA61/NA49 meeting, Wrocław

Reference cloud cluster The virtual machines require a cluster of physical hosts A reference cloud cluster has been created Detailed documentation will simplify the process of replicating it at other sites Based on OpenNebula (popular cloud framework) KVM hypervisor Provides Amazon EC2 interface (de facto standard for cloud management) 05.03.2012 NA61/NA49 meeting, Wrocław

Data bookkeeping web interface A web interface for bookkeeping of the data to be created List all existing data with status (e.g. software versions used for processing) Easy selection of data for (re)processing with selected OS and software version A virtual on-demand cluster is created After processing, data written back to Castor Either based on existing frameworks, or on new development Likely using EC2 interface for the cloud management Can allow for great flexibility of processing site 05.03.2012 NA61/NA49 meeting, Wrocław

Roadmap Task Status/done Remaining Expected NA61 software installation Legacy framework SHINE End of March? NA49 software installation Development machine, software repository Software installation Facilitate batch system Condor job scripts Modifications/bug fixes March Validate outputs Small data set Large data set (using batch system) End of March On-demand virtual cluster Cluster creation / destroy scripts Production reconstruction Dependencies mostly ready Remaining tasks, prepare for real reconstruction April Reference cloud cluster Cluster working Documentation June/July Data bookkeeping web interface Initial planing Evaluate frameworks “First” version “Final” version End of October 05.03.2012 NA61/NA49 meeting, Wrocław

Input needed NA49 software installation Eventual SHINE issues Eventual validation issues How to practically arrange for production reconstruction Please keep virtualisation (CERNVM/CVMFS) in mind when making plans ... 05.03.2012 NA61/NA49 meeting, Wrocław