Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

National Grid's Contribution to LHCb IFIN-HH Serban Constantinescu, Ciubancan Mihai, Teodor Ivanoaica.
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site Adam Padee ( ) Ryszard Gokieli ( ) Krzysztof.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
A comparison between xen and kvm Andrea Chierici Riccardo Veraldi INFN-CNAF.
HS06 on the last generation of CPU for HEP server farm Michele Michelotto 1.
UCL Site Report Ben Waugh HepSysMan, 22 May 2007.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Testing Virtual Machine Performance Running ATLAS Software Yushu Yao Paolo Calafiura LBNL April 15,
RAL PPD Site Update and other odds and ends Chris Brew.
Planning and Designing Server Virtualisation.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
Farm Completion Beat Jost and Niko Neufeld LHCb Week St. Petersburg June 2010.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Tier-1 Batch System Report Andrew Lahiff, Alastair Dewhurst, John Kelly, Ian Collier 5 June 2013, HEP SYSMAN.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
3. April 2006Bernd Panzer-Steindel, CERN/IT1 HEPIX 2006 CPU technology session some ‘random walk’
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
Multi-core jobs at the RAL Tier-1 Andrew Lahiff, Alastair Dewhurst, John Kelly February 25 th 2014.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
An Agile Service Deployment Framework and its Application Quattor System Management Tool and HyperV Virtualisation applied to CASTOR Hierarchical Storage.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
Grid DESY Andreas Gellrich DESY EGEE ROC DECH Meeting FZ Karlsruhe, 22./
1 Evolving ATLAS Computing Model and Requirements Michael Ernst, BNL With slides from Borut Kersevan and Karsten Koeneke U.S. ATLAS Distributed Facilities.
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
Arne Wiebalck -- VM Performance: I/O
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
Ian Gable HEPiX Spring 2009, Umeå 1 VM CPU Benchmarking the HEPiX Way Manfred Alef, Ian Gable FZK Karlsruhe University of Victoria May 28, 2009.
Workload management, virtualisation, clouds & multicore Andrew Lahiff.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
PROOF Benchmark on Different Hardware Configurations 1 11/29/2007 Neng Xu, University of Wisconsin-Madison Mengmeng Chen, Annabelle Leung, Bruce Mellado,
Maria Girone CERN - IT Tier0 plans and security and backup policy proposals Maria Girone, CERN IT-PSS.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
Atlas Software Structure Complicated system maintained at CERN – Framework for Monte Carlo and real data (Athena) MC data generation, simulation and reconstruction.
UKI-LT2-RHUL ATLAS STEP09 report Duncan Rand on behalf of the RHUL Grid team.
Scientific Computing in PPD and other odds and ends Chris Brew.
RALPP Site Report HEP Sys Man, 11 th May 2012 Rob Harper.
1 Cloud Services Requirements and Challenges of Large International User Groups Laurence Field IT/SDC 2/12/2014.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Virtual Server Server Self Service Center (S3C) JI July.
Alessandro De Salvo CCR Workshop, ATLAS Computing Alessandro De Salvo CCR Workshop,
Performance analysis comparison Andrea Chierici Virtualization tutorial Catania 1-3 dicember 2010.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Using HLRmon for advanced visualization of resource usage Enrico Fattibene INFN - CNAF ISCG 2010 – Taipei March 11 th, 2010.
Atlas Tier 3 Overview Doug Benjamin Duke University.
Using the CMS Higher Level Trigger Farm as a Cloud Resource David Colling Imperial College London.
CMS Multicore jobs at RAL Andrew Lahiff, RAL WLCG Multicore TF Meeting 1 st July 2014.
Multi-Core CPUs Matt Kuehn. Roadmap ► Intel vs AMD ► Early multi-core processors ► Threads vs Physical Cores ► Multithreading and Multi-core processing.
Lessons learned administering a larger setup for LHCb
Getting the Most out of Scientific Computing Resources
Title of the Poster Supervised By: Prof.*********
Experience of Lustre at QMUL
Getting the Most out of Scientific Computing Resources
Cluster Optimisation using Cgroups
WLCG Memory Requirements
Experience of Lustre at a Tier-2 site
UK GridPP Tier-1/A Centre at CLRC
Oxford Site Report HEPSYSMAN
Статус ГРИД-кластера ИЯФ СО РАН.
Presentation transcript:

Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using Hyper-Threading RAL Tier1 Setup The RAL TIER1 batch farm consists of several multicore, hyperthreading capable CPUs. Increases in the amount of memory per node combined with experiences from other sites made hyperthreading an attractive option for increasing job throughput. RAL supports all LHC VOs, with prime users being Atlas, CMS and LHCB, and a 10% of resources is devoted to non-LHC VOs The virtual cores provided by hyperthreading could double the batch farm capacity, however the amount of memory available in the batch nodes did not permit that. LHC jobs require more than 2GB RAM to run smoothly. With all HT cores enabled total job slots capacity could double. In practice memory constrains resulted in an increase of 30% RAL Tier1 Setup The RAL TIER1 batch farm consists of several multicore, hyperthreading capable CPUs. Increases in the amount of memory per node combined with experiences from other sites made hyperthreading an attractive option for increasing job throughput. RAL supports all LHC VOs, with prime users being Atlas, CMS and LHCB, and a 10% of resources is devoted to non-LHC VOs The virtual cores provided by hyperthreading could double the batch farm capacity, however the amount of memory available in the batch nodes did not permit that. LHC jobs require more than 2GB RAM to run smoothly. With all HT cores enabled total job slots capacity could double. In practice memory constrains resulted in an increase of 30% Method Each generation of batch farm hardware with hyperthreading capability was benchmarked with HEPSPEC, progressively increasing the number of threads up to the total number of virtual cores Benchmarks at that time were conducted using Scientific Linux 5. Scientific Linux 6 benchmarks were run later as the batch farm was set to be updated. Scientific Linux 6 performed slightly better than Scientific Linux 5. The overall trend was identical Power, temperature, disk I/O and batch server performance were closely monitored The results indicated a nearly linear increase in the hepspec scores, flattening at about 14 threads for 2 CPU 4 core nodes and 20 threads for 2 CPU 6 core nodes The revealed sweet spots were then configured for use in the batch farm to discover where production VO jobs would perform optimally Method Each generation of batch farm hardware with hyperthreading capability was benchmarked with HEPSPEC, progressively increasing the number of threads up to the total number of virtual cores Benchmarks at that time were conducted using Scientific Linux 5. Scientific Linux 6 benchmarks were run later as the batch farm was set to be updated. Scientific Linux 6 performed slightly better than Scientific Linux 5. The overall trend was identical Power, temperature, disk I/O and batch server performance were closely monitored The results indicated a nearly linear increase in the hepspec scores, flattening at about 14 threads for 2 CPU 4 core nodes and 20 threads for 2 CPU 6 core nodes The revealed sweet spots were then configured for use in the batch farm to discover where production VO jobs would perform optimally Results Overall, 2000 extra job slots and 9298 extra hepspec were added in the batch farm using already available hardware Average job time increases as expected, but overall job throughput increased Network/disk/power/temperature usage did not increase in a way that could negatively affect the overall throughput or require additional maneuvers Batch server was able to handle the extra job slots Of critical importance is the sharp drop in job efficiency as job slots approach the upper hyperthreading limit. This means that real world VO jobs would suffer if we went for full batchfarm HEPSPEC performance! Results Overall, 2000 extra job slots and 9298 extra hepspec were added in the batch farm using already available hardware Average job time increases as expected, but overall job throughput increased Network/disk/power/temperature usage did not increase in a way that could negatively affect the overall throughput or require additional maneuvers Batch server was able to handle the extra job slots Of critical importance is the sharp drop in job efficiency as job slots approach the upper hyperthreading limit. This means that real world VO jobs would suffer if we went for full batchfarm HEPSPEC performance! Conclusions New procurements now take into account the hyperthreading capabilities For 2012, dual 8 core CPU systems go up to 32 virtual cores Systems were procured with 128 GB RAM in order to exploit full hyperthreading capabilities Dual Gigabit links, in the future single 10 GB as they became more cost effective So far RAID0 software raid setup has proven sufficient for disk I/O Performance gains so far on par with previous generations By spending a bit extra on RAM, we save more by buying fewer nodes This also saves machine room space, cables, and power Conclusions New procurements now take into account the hyperthreading capabilities For 2012, dual 8 core CPU systems go up to 32 virtual cores Systems were procured with 128 GB RAM in order to exploit full hyperthreading capabilities Dual Gigabit links, in the future single 10 GB as they became more cost effective So far RAID0 software raid setup has proven sufficient for disk I/O Performance gains so far on par with previous generations By spending a bit extra on RAM, we save more by buying fewer nodes This also saves machine room space, cables, and power MakeJob Slots per WNEfficiencyAverage Job Length (mins)Standard Deviation (mins)Number of jobs Dell Viglen Dell Viglen Dell Viglen Dell Viglen Dell Viglen Dell Viglen Dell Viglen