NA61/NA49 virtualisation: status and plans Dag Toppe Larsen Budapest 14.05.2012.

Slides:



Advertisements
Similar presentations
1. Topics Is Cloud Computing the way to go? ARC ABM Review Configuration Basics Setting up the ARC Cloud-Based ABM Hardware Configuration Software Configuration.
Advertisements

Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
New Cluster for Heidelberg TRD(?) group. New Cluster OS : Scientific Linux 3.06 (except for alice-n5) Batch processing system : pbs (any advantage rather.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Reproducible Environment for Scientific Applications (Lab session) Tak-Lon (Stephen) Wu.
To run the program: To run the program: You need the OS: You need the OS:
Space Science and Engineering Center University of Wisconsin-Madison Virtual Machines: A method for distributing DB processing software Liam Gumley.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Copyright © 2010 Platform Computing Corporation. All Rights Reserved.1 The CERN Cloud Computing Project William Lu, Ph.D. Platform Computing.
DIRAC Web User Interface A.Casajus (Universitat de Barcelona) M.Sapunov (CPPM Marseille) On behalf of the LHCb DIRAC Team.
Testing Virtual Machine Performance Running ATLAS Software Yushu Yao Paolo Calafiura LBNL April 15,
1 port BOSS on Wenjing Wu (IHEP-CC)
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Introduction to CVMFS A way to distribute HEP software on cloud Tian Yan (IHEP Computing Center, BESIIICGEM Cloud Computing Summer School.
Update on Database Issues Peter Chochula DCS Workshop, June 21, 2004 Colmar.
Data production using CernVM and lxCloud Dag Toppe Larsen Belgrade
Predrag Buncic (CERN/PH-SFT) CernVM - a virtual software appliance for LHC applications C. Aguado-Sanchez 1), P. Buncic 1), L. Franco 1), A. Harutyunyan.
Status of StoRM+Lustre and Multi-VO Support YAN Tian Distributed Computing Group Meeting Oct. 14, 2014.
Data production using CernVM and LxCloud Dag Toppe Larsen Warsaw,
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
Predrag Buncic (CERN/PH-SFT) WP9 - Workshop Summary
Installing, running, and maintaining large Linux Clusters at CERN Thorsten Kleinwort CERN-IT/FIO CHEP
Virtualised Worker Nodes Where are we? What next? Tony Cass GDB /12/12.
Infrastructure for QA and automatic trending F. Bellini, M. Germain ALICE Offline Week, 19 th November 2014.
DPHEP Workshop CERN, December Predrag Buncic (CERN/PH-SFT) CernVM R&D Project Portable Analysis Environments using Virtualization.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
GLIDEINWMS - PARAG MHASHILKAR Department Meeting, August 07, 2013.
WLCG Overview Board, September 3 rd 2010 P. Mato, P.Buncic Use of multi-core and virtualization technologies.
Portal Update Plan Ashok Adiga (512)
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
Maite Barroso - 10/05/01 - n° 1 WP4 PM9 Deliverable Presentation: Interim Installation System Configuration Management Prototype
Predrag Buncic (CERN/PH-SFT) Virtualizing LHC Applications.
Condor + Cloud Scheduler Ashok Agarwal, Patrick Armstrong, Andre Charbonneau, Ryan Enge, Kyle Fransham, Colin Leavett-Brown, Michael Paterson, Randall.
Status of BESIII Distributed Computing BESIII Workshop, Sep 2014 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Selenium server By, Kartikeya Rastogi Mayur Sapre Mosheca. R
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
NA61 Collaboration Meeting CERN, December Predrag Buncic, Mihajlo Mudrinic CERN/PH-SFT Enabling long term data preservation.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
EGI-InSPIRE RI EGI Webinar EGI-InSPIRE RI Porting your application to the EGI Federated Cloud 17 Feb
Software framework and batch computing Jochen Markert.
Virtual Server Server Self Service Center (S3C) JI July.
© ExplorNet’s Centers for Quality Teaching and Learning 1 Explain the purpose of Microsoft virtualization. Objective Course Weight 2%
Advancing CernVM-FS and its Development Infrastructure José Molina Colmenero CERN EP-SFT.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Virtualisation: status and plans Dag Toppe Larsen
Predrag Buncic, CERN/PH-SFT The Future of CernVM.
HEPiX Virtualisation working group Andrea Chierici INFN-CNAF Workshop CCR 2010.
Use of HLT farm and Clouds in ALICE
Virtualisation for NA49/NA61
NA61/NA49 virtualisation:
Blueprint of Persistent Infrastructure as a Service
Dag Toppe Larsen UiB/CERN CERN,
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Dag Toppe Larsen UiB/CERN CERN,
ATLAS Cloud Operations
Introduction to CVMFS A way to distribute HEP software on cloud
Virtualisation for NA49/NA61
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
WLCG Collaboration Workshop;
Presentation transcript:

NA61/NA49 virtualisation: status and plans Dag Toppe Larsen Budapest

NA61/NA49 meeting, Budapest2 Outline Quick reminder of CERNVM Tasks Each task in detail Roadmap CernVM installation Input needed

NA61/NA49 meeting, Budapest3 CERNVM CERNVM is a Linux-distribution Designed specifically for virtual machines (VMs) Based on SLC (currently SLC5) Compressed image size ~300MB Both 32-bit and 64-bit versions Addition software “Standard” software via Conary package manager Experiment software via CVMFS Contextualisation: images adapted to experiment requirements during boot Data preservation: all images are permanently preserved

NA61/NA49 meeting, Budapest4 CVMFS Distributed read-only file system for CERNVM (i.e. the same as AFS for LXPLUS) Can also be used by “real” machines (e.g. LXPLUS, grid) Files compressed and distributed via HTTP  Global availability Central server, site replication via standard HTTP proxies Files decompressed and cached on (CERNVM) computer  Can run without Internet access if all needed files are cached Mainly for experimental software, but also other “static” data (e.g. calibration data) Each experiment has a repository to store all versions of software Common software (e.g. ROOT) available from SFT repository

NA61/NA49 meeting, Budapest5 Data preservation As technology evolves, no longer possible to run legacy software on modern platforms Must be preserved and accessible: Experiment data Experiment software Operating environment (operating system, libraries, compilers, hardware) Just preserving data and software is not enough Virtualisation may preserve operating environment

NA61/NA49 meeting, Budapest6 CERNVM data preservation “Solution”: Experiment data stored on Castor Experiment software versions stored on CVMFS  HTTP “lasting” technology Operation environments stored as CERNVM image versions Thus, a legacy version of CERNVM can be started as a VM, running a legacy version of experiment software Forward-looking approach (we start preserving now)

NA61/NA49 meeting, Budapest7 Tasks Make experiment software available Facilitate batch processing Validate outputs On-demand virtual clusters Production reconstruction Reference cloud cluster Data bookkeeping web interface

NA61/NA49 meeting, Budapest8 Make experiment software available NA61/NA49 software must be available on CVMFS for CernVM to process data NA61 Legacy software chain installed  Changes to be fed back to SVN SHINE  Experts are installing production version on CVMFS  CernVM environment has to be made  Automatic install of necessary packages via Conary  SVN checkout should compile “out of the box”  May be better to use 32-bit CernVM image NA49 Software has been installed (validation needed)

NA61/NA49 meeting, Budapest9 Facilitate batch processing LXPLUS uses PBS batch system CernVM uses Condor batch system “Philosophical” differences PBS has one job script per job Condor has common job description file with parameters for each job Existing PBS scripts have been ported to Condor

NA61/NA49 meeting, Budapest10 Output validation – status Run 8688 has been processed on both CernVM/CVMFS and LXPLUS/AFS, using software version v2r7g According to analysis by Grzegorz, there are relatively large discrepancies (larger than SLC4->5) Surprising, since software is the same, and CernVM IS Scientific Linux 5 (just repacked) Can there be issues with calibration files? Or some of the changes done to make software work on CVMFS?

NA61/NA49 meeting, Budapest11 Output validation – plan Have requested new reconstruction on LXPLUS/CVMFS Will make it possible to separate effect of CernVM/LXPLUS from CVMFS/AFS (three-way comparison) Shine is production-ready now Should we “forget” legacy chain, and focus on Shine? On the other hand: NA49 reconstruction may have same discrepancy as NA61 legacy chain; good reason to still investigate source

NA61/NA49 meeting, Budapest12 On-demand virtual clusters A cluster may need VMs of different configurations, depending on type of jobs Memory, CernVM version, experiment SW, etc. Thus, need for dynamic creation/destruction of virtual cluster Created command-line script for creating virtual clusters Later to be controlled by data bookkeeping web interface

NA61/NA49 meeting, Budapest13 Production reconstruction Production reconstruction next step Cluster of “decent” size needed Need to submit ~50 VMs to process a large data set To run on LXCLOUD (experimental CERN service) Awaiting conclusion from software validation step

NA61/NA49 meeting, Budapest14 Reference cloud cluster The virtual machines require a cluster of physical hosts A reference cloud cluster has been created Detailed documentation will simplify the process of replicating it at other sites Based on OpenNebula (popular cloud framework) KVM hypervisor Provides Amazon EC2 interface (de facto standard for cloud management)

NA61/NA49 meeting, Budapest15 Data bookkeeping web interface A web interface for bookkeeping of the data to be created List all existing data with status (e.g. software versions used for processing) Easy selection of data for (re)processing with selected OS and software version A virtual on-demand cluster is created After processing, data written back to Castor Either based on existing frameworks, or on new development Using EC2 interface for the cloud management Allows for great flexibility of processing site

NA61/NA49 meeting, Budapest16 Bookkeeping outlook Most important/urgent task Founding ends end of October All dependencies (software, cloud, dynamic clusters) in place Only bookkeeping system itself missing Optimistic about outlook for completion by end of October But should be wary of sources (validation?) of delays...

NA61/NA49 meeting, Budapest17 Roadmap TaskStatus/doneRemainingExpected NA61 software installation Legacy frameworkShine frameworkEnd of May NA49 software installation Software installedData validationTo be determined Facilitate batch system OK Validate outputsCreated reference data sets for validation Understand source of discrepancies End of May On-demand virtual cluster OK Production reconstruction Set-up readyAwaiting conclusion of validation discrepancies End of May Reference cloud cluster Cluster workingDocumentationEnd of July Data bookkeeping web interface Initial planingEvaluate frameworks “First” version “Final” version End of May End of July End of October

NA61/NA49 meeting, Budapest18 Next steps Parallel task 1 Understand validation discrepancies Run large-scale processing on CernVM Parallel task 2 (critical path) Data bookkeeping web interface for CernVM processing Run large-scale processing using CernVM/web interface Transfer to NA61

NA61/NA49 meeting, Budapest19 CernVM for development CernVM makes it possible to run production version of legacy software/shine on laptop without local install Also possible to compile Shine from SVN on CernVM “out of the box” when the proper NA61 environment is set up Is also possible to mount NA61 software from CVMFS directly on laptop (but software dependencies may have to be resolved by user)

NA61/NA49 meeting, Budapest20 CernVM installation on laptop Install a hypervisor of your choice, e.g. Virtualbox: Download a matching CernVM desktop image: Open :8004 in your web browser (user=admin, password=password) Select NA61 and PH-SFT software repositories Reboot You are now ready to use NA61 software in CernVM on your laptop! More information:

NA61/NA49 meeting, Budapest21 My schedule I will be around for assistance with installing CernVM on laptops However, I plan to leave Wednesday morning, since I have a CHEP-poster to print/make sure is brought to CHEP (I will not go to CHEP, only my poster will) This is “negotiable”, e.g. if there is big demand for CernVM install, Shine installation work or important discussions

NA61/NA49 meeting, Budapest22 Input needed NA49 validation SHINE installation NA61 legacy validation discrepancies How to practically arrange for large scale reconstruction Issues related to data bookkeeping Please keep virtualisation (CERNVM/CVMFS) in mind when making plans...