Oxford Site Update HEPiX Sean Brisbane Tier 3 Linux System Administrator March 2015.

Slides:



Advertisements
Similar presentations
UCL HEP Computing Status HEPSYSMAN, RAL,
Advertisements

Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
Birmingham site report Lawrie Lowe: System Manager Yves Coppens: SouthGrid support HEP System Managers’ Meeting, RAL, May 2007.
17th October 2013Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
IHEP Site Status Jingyan Shi, Computing Center, IHEP 2015 Spring HEPiX Workshop.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Microsoft virtual machine converter
RAL Site Report HEPiX Fall 2013, Ann Arbor, MI 28 Oct – 1 Nov Martin Bly, STFC-RAL.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
DMF Configuration for JCU HPC Dr. Wayne Mallett Systems Manager James Cook University.
A. Mohapatra, HEPiX 2013 Ann Arbor1 UW Madison CMS T2 site report D. Bradley, T. Sarangi, S. Dasu, A. Mohapatra HEP Computing Group Outline  Infrastructure.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
SouthGrid Status Pete Gronbech: 2 nd April 2009 GridPP22 UCL.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPIX 2009 Umea, Sweden 26 th May 2009.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN RAL 30 th June 2009.
HTCondor and Beyond: Research Computer at Syracuse University Eric Sedore ACIO – Information Technology and Services.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
Configuration Management with Cobbler and Puppet Kashif Mohammad University of Oxford.
RAL PPD Computing A tier 2, a tier 3 and a load of other stuff Rob Harper, June 2011.
Lucien Boland and Sean Crosby Research Computing.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
NCPHEP ATLAS/CMS Tier3: status update V.Mossolov, S.Yanush, Dz.Yermak National Centre of Particle and High Energy Physics of Belarusian State University.
13th October 2011Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
RAL Site Report Castor Face-to-Face meeting September 2014 Rob Appleyard, Shaun de Witt, Juan Sierra.
RAL Site Report HEPiX FAll 2014 Lincoln, Nebraska October 2014 Martin Bly, STFC-RAL.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Oxford & SouthGrid Update HEPiX Pete Gronbech GridPP Project Manager October 2015.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
11th October 2012Graduate Lectures1 Oxford University Particle Physics Unix Overview Pete Gronbech Senior Systems Manager and GridPP Project Manager.
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
BaBar Cluster Had been unstable mainly because of failing disks Very few (
R. Krempaska, October, 2013 Wir schaffen Wissen – heute für morgen Controls Security at PSI Current Status R. Krempaska, A. Bertrand, C. Higgs, R. Kapeller,
RALPP Site Report HEP Sys Man, 11 th May 2012 Rob Harper.
1 Update at RAL and in the Quattor community Ian Collier - RAL Tier1 HEPiX FAll 2010, Cornell.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
CernVM-FS Infrastructure for EGI VOs Catalin Condurache - STFC RAL Tier1 EGI Webinar, 5 September 2013.
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
The Great Migration: From Pacman to RPMs Alain Roy OSG Software Coordinator.
DPM: Future Proof Storage Ricardo Rocha ( on behalf of the DPM team ) EMI INFSO-RI
EMI is partially funded by the European Commission under Grant Agreement RI Future Proof Storage with DPM Oliver Keeble (on behalf of the CERN IT-GT-DMS.
UK Status and Plans Catalin Condurache – STFC RAL ALICE Tier-1/Tier-2 Workshop University of Torino, February 2015.
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
Experience of Lustre at QMUL
Pete Gronbech GridPP Project Manager April 2016
HEPiX Spring 2014 Annecy-le Vieux May Martin Bly, STFC-RAL
Experience of Lustre at a Tier-2 site
Oxford Site Report HEPSYSMAN
BEIJING-LCG2 Site Report
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
ETHZ, Zürich September 1st , 2016
Upgrading Your Private Cloud with Windows Server 2012 R2
This work is supported by projects Research infrastructure CERN (CERN-CZ, LM ) and OP RDE CERN Computing (CZ /0.0/0.0/1 6013/ ) from.
RHUL Site Report Govind Songara, Antonio Perez,
Pete Gronbech, Kashif Mohammad and Vipul Davda
Kashif Mohammad VIPUL DAVDA
Presentation transcript:

Oxford Site Update HEPiX Sean Brisbane Tier 3 Linux System Administrator March 2015

Contents The Oxford Physics department Working environments for HEP The Oxford Tier 2 and 3 Experience with Lustre 2.5 Experience with puppet University upgrades Oxford – March

Oxford Particle Physics - Overview Oxford University has one of the largest Physics Departments in the UK –~450 staff and 1100 students. Particle Physics sub-department is the largest in the UK Department now supports Windows, Ubuntu, Mac on the desktop across the department –Ubuntu desktops and departmental Linux infrastructure sharing some staff with HEP –Cobbler used to install base system, Cfengine2 used for package control and configuration. Two SL6 computational clusters for the HEP –Grid Cluster part of the SouthGrid Tier-2 –Local Cluster (AKA Tier-3) Cobbler and Puppet for all SL6 machines –Additional SL6 desktops (HEP only) off the back of puppet system –Significant investment in a clean module, node and configuration hierarchy Oxford – March

Oxford Cluster’s Hardware Most of the storage now Dell R510 / R720xd –TB 1,300 (T2) and 700 TB (T3) –Most recently used 12x SATA not 12x SAS this time. Unlikely to do this in the future. –SAS becoming the default and support for SATA costs extra. –Disk I/O on T3 benefits from SAS –Raid cards ?possibly? rejecting SATA Most of the compute is supermicro twin squared –HS06 16,800 T2 and 7,200 T3 Most T3 infrastructure nodes VMWare and most T2 OVirt Oxford – March

Oxford’s Grid Cluster 5 Capacity:- 16,800 HS06, 1,300TB

Grid Cluster Two thirds HT Condor behind ARC CE. –Soft cgroup limit on CPU and memory Remaining third running legacy torque/maui driven by CREAM CE. Storage is DPM Small test beds of alternative worker technologies –VAC –Open Stack Oxford runs the Nagios based WLCG monitoring for the UK –These include the Nagios server itself, and support nodes for it, SE, MyProxy and WMS/LB –Multi VO Nagios Monitoring added two years ago. Take a leading role in IPv6 Testing –Our campus network is barely IPv6 capable but will be replaced over the coming year Oxford – March

Local Computer room showing PP cluster & cold aisle containment 7 Very similar h/w to the Grid Cluster. Same Cobbler and Puppet management setup. Lustre used for larger groups Capacity: HS06, 716TB

Local PP cluster 8 Capacity: HS06, 720TB –LHCb and Atlas use Lustre (~550TB) –NFS (cdf, t2k, sno+, dark matter, accelerator science, mars) SL6 + torque –Small <5% SL5 (or ubuntu) requirement met with a chroot on SL6 –Need to translate maui configs blocking move to HT condor All remote nodes –Connect with ssh + X windows or xrdp –T3 disks available to department over SMB, NFSv4 (krb5) and webdav interfaces CVMFS stratum zero for local software –Allows external sharing to university and beyond –Top-level separation between software for generic Linux, Ubuntu and SL versions Oxford – March 2015

Lustre Our Lustre experience in summary Lustre 1.8 worked really well. Lustre 2.1 we had no problems but didn’t try to decommission or move any storage nodes Lustre 2.5 has been more problematic than anticipated Lustre 2.5 Our lustre 2.5 life history –Start with a clean lustre 2.5 filesystem (ldiskfs) –Retired older storage server/targets using lfs_migrate on 2.5 file-system –Migrated metadata server (mds) from old to new hardware within the new 2.5 file system using tar/untar This process has now disappeared from the l2.x manual Some number of annoying issues –Documentation outdated –Some operations no longer supported as they were in lustre 1.8 –File system is overall slower Some number of more serious bugs resulting in corruption and loss of file-system parallelism Oxford – March

A selection of our lustre issues The following are traced to a Lustre bug (we think) 1.lfs find misbehaves for some OST names if OST name rather than index used [1] –Just use find by index 2.Large directory corruption not fully categorized on mds [2] –“..” (dot dot) Entry in wrong place. –Note for those running lustre there is a different related bug. –Htree conversion messed up for large directories 3.Cannot free space on deactivated OST –Just don’t deactivate when freeing space [1] possibly LU-1738 [2] LU5626, [3] LU-4825 Not understood OST round robin and space-aware placements stopped working (not reported ) –Assume decommissioning OST triggers the edge cases? –Makes us want [3] in production - for manual rebalance –Accumulation of files on a few OSTs All compounded by out of date documentation Oxford – March

Integrate with university cluster? Becoming more attractive for local and grid use –Recent upgrade to 6000 cores for university RHEL6 cluster – free at point of use –Campus network upgrade (~6 months) may give us a 10G link in to the service How to configure working environment to reduce barrier to entry for users? With cvmfs –Challenges getting HEP software installed (cvmfs + fuse) as we scale up or try to run grid jobs –Per-user self configuration with parrot works at low scale Data access –HPC facility panasas storage is cluster-local and limited to 5TB/group –Looking at Xrootd + Kerberos / X509 to our T3 disks / the grid Some challenges reconciling our change management policies/expectations Current status Local HEP users offload bulk, standalone, low data rate jobs to this cluster Keep other jobs on T3 Oxford – March 20standalone 15 11

Oxford – March Summary, Questions? Challenges integrating multiple client OS Two solid computer rooms Involvement in many grid development projects Issues with Lustre 2.5 Puppet is powerful, but needs an investment of time to get right Involving University HPC clusters

Additional Material Oxford – March

Puppet Separate cfengine2 each for SL =10.04 SL6 straight to puppet A lot of effort has gone into our puppet design Puppet is a framework from which to build a configuration management system Collaboration on shared puppet modules is fantastic but not the whole story Categorize your site and build a model (hierarchy) When all this is done, one has a clean and flexible infrastructure Balancing flexibility, duplication (i.e. forgetting to change the oddball), complexity of the module lists and configs Oxford – March

Puppet Hierarchy Oxford – March 2015 Nodes + hiera configs share a hierarchy Grid Local client Common Desktop WN Worker DPM t2wn1 t2wn2 pplxdt0 1 Common Cluster Role Node abstractionsfunctional staticclientNFS clientsssd PP software cvmfs Yum packages Modules