S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment S.Chechelnitskiy Simon Fraser University CHEP 2007 September 6 th.

Slides:



Advertisements
Similar presentations
Express5800/ft series servers Product Information Fault-Tolerant General Purpose Servers.
Advertisements

Virtualisation From the Bottom Up From storage to application.
EGEE is a project funded by the European Union under contract IST Using SRM: DPM and dCache G.Donvito,V.Spinoso INFN Bari
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
6/2/20071 Grid Computing Sun Grid Engine (SGE) Manoj Katwal.
1/28/2004CSCI 315 Operating Systems Design1 Operating System Structures & Processes Notice: The slides for this lecture have been largely based on those.
Cloud Computing and Virtualization Sorav Bansal CloudCamp 2010 IIT Delhi.
VIRTUALISATION OF HADOOP CLUSTERS Dr G Sudha Sadasivam Assistant Professor Department of CSE PSGCT.
Scalable Server Load Balancing Inside Data Centers Dana Butnariu Princeton University Computer Science Department July – September 2010 Joint work with.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
VAP What is a Virtual Application ? A virtual application is an application that has been optimized to run on virtual infrastructure. The application software.
SANPoint Foundation Suite HA Robert Soderbery Sr. Director, Product Management VERITAS Software Corporation.
SSI-OSCAR A Single System Image for OSCAR Clusters Geoffroy Vallée INRIA – PARIS project team COSET-1 June 26th, 2004.
A Cloud is a type of parallel and distributed system consisting of a collection of inter- connected and virtualized computers that are dynamically provisioned.
OPERATING SYSTEMS CPU SCHEDULING.  Introduction to CPU scheduling Introduction to CPU scheduling  Dispatcher Dispatcher  Terms used in CPU scheduling.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Grid Canada CLS eScience Workshop 21 st November, 2005.
So, Jung-ki Distributed Computing System LAB School of Computer Science and Engineering Seoul National University Implementation of Package Management.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks C. Loomis (CNRS/LAL) M.-E. Bégin (SixSq.
:: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: :: GridKA School 2009 MPI on Grids 1 MPI On Grids September 3 rd, GridKA School 2009.
J. J. Rehr & R.C. Albers Rev. Mod. Phys. 72, 621 (2000) A “cluster to cloud” story: Naturally parallel Each CPU calculates a few points in the energy grid.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Sensitivity of Cluster File System Access to I/O Server Selection A. Apon, P. Wolinski, and G. Amerson University of Arkansas.
First experience of submission to the EGEE/RDIG Grid of jobs prepared for non standart OSs by means of virtualization. I.Gorbunov, A.Kryukov SINP MSU,
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
High Performance Computing on Virtualized Environments Ganesh Thiagarajan Fall 2014 Instructor: Yuzhe(Richard) Tang Syracuse University.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Using Virtual Servers for the CERN Windows infrastructure Emmanuel Ormancey, Alberto Pace CERN, Information Technology Department.
Ashok Agarwal University of Victoria 1 GridX1 : A Canadian Particle Physics Grid A. Agarwal, M. Ahmed, B.L. Caron, A. Dimopoulos, L.S. Groer, R. Haria,
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Headline in Arial Bold 30pt HPC User Forum, April 2008 John Hesterberg HPC OS Directions and Requirements.
1 CMPE 511 HIGH PERFORMANCE COMPUTING CLUSTERS Dilek Demirel İşçi.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Tools and techniques for managing virtual machine images Andreas.
| nectar.org.au NECTAR TRAINING Module 4 From PC To Cloud or HPC.
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
SL5 Site Status GDB, September 2009 John Gordon. LCG SL5 Site Status ASGC T1 - will be finished before mid September. Actually the OS migration process.
Desktop Workload Characterization for CMP/SMT and Implications for Operating System Design Sven Bachthaler Fernando Belli Alexandra Fedorova Simon Fraser.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
2007/05/22 Integration of virtualization software Pierre Girard ATLAS 3T1 Meeting
© 2010 VMware Inc. All rights reserved Why Virtualize? Beng-Hong Lim, VMware, Inc.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Chapter 6: Securing the Cloud
The EDG Testbed Deployment Details
The Beijing Tier 2: status and plans
Dag Toppe Larsen UiB/CERN CERN,
Dag Toppe Larsen UiB/CERN CERN,
GWE Core Grid Wizard Enterprise (
Database Services at CERN Status Update
Clinton A Jones Eastern Kentucky University Department of Technology
BDII Performance Tests
Andrew McCombs March 10th, 2011
Quattor Usage at Nikhef
Oracle Solaris Zones Study Purpose Only
Introduction to Cloud Computing
CPU SCHEDULING.
New cluster capabilities through Checkpoint/ Restart
Presentation transcript:

S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment S.Chechelnitskiy Simon Fraser University CHEP 2007 September 6 th 2007 ATLAS on

S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment Motivation Proposed architecture Required software XEN performance tests Test suite and results Implementation schedule Running SE in XEN Conclusion ATLAS on

S.Chechelnitskiy / SFU Simon Fraser Motivation SFU is a Tier-2 site SFU has to serve its researchers Funding is done through WestGrid only WestGrid is a consortia of Canadian Universities. It operates a high performance computing (HPC), collaboration and visualization infrastructure across Western Canada. WestGrid SFU cluster must be an Atlas Tier-2 CE. CE must run on a WestGrid facility

S.Chechelnitskiy / SFU Simon Fraser Motivation (cont.) Current WestGrid SFU cluster load 70% of jobs are serial 30% of jobs are parallel (MPICH-2) 10% of serial jobs are preemptible Scheduling policies No walltime limit for serial jobs 1 MPI job per user with Ncpu = 1/4 of Ncpu_total Minimize waiting time through the use of preemption Extensive backfill policies

S.Chechelnitskiy / SFU Simon Fraser Motivation (cont.) Atlas and local jobs must run on the same hardware. Atlas Cluster Requirements: Particular operating system LCG software (+experimental software, updates, etc.) Connectivity to the Internet WestGrid (local) Cluster Requirements: No connectivity to the Internet Lots of different software Recent operating systems Low latency for MPI jobs

S.Chechelnitskiy / SFU Simon Fraser Proposed solution 3 Different Environments for 3 Jobs Types: WNs run a recent operating system with XEN capability (openSUSe-10.2) MPI jobs run in the non-virtualized environment For each serial job (Atlas or local) a separate virtual machine (XEN) is started on a WN (max number of XENS == Number of cores) Local serial jobs start XEN with openSUSe-10.2 and local collection of software Atlas jobs start XEN with SC-4 and LCG software 2 separate hardware headnodes for the local and Atlas clusters

S.Chechelnitskiy / SFU Simon Fraser Proposed architecture

S.Chechelnitskiy / SFU Simon Fraser Software Requirements OS with XEN capability Recent Torque version >= 2.0 Moab cluster manager, version >= 5.0 Some modifications to LCG software is necessary

S.Chechelnitskiy / SFU Simon Fraser XEN Performance Tests Performance for serial jobs in XEN is better than 95% of the performance of the same job running in a non-virtualized environment and the same OS Network bandwidth is still a problem for XEN but it is not required by serial jobs (neither Atlas, nor local) Memory usage for one XEN virtual machine is about MB XEN startup time is less than a minute

S.Chechelnitskiy / SFU Simon Fraser Test Suite and Results Test Suite: One SUN X2100 as an Atlas headnode, software installed: LCG-CE, Moab (master), mounts /var/spool/pbs from the local headnode for accounting purposes One SUN V20Z as a local headnode, software installed: Moab (slave), Torque Two SUN X4100 (dual core) as WNs, software installed: torque_mom, XEN-kernel Images repository is located on the headnodes and mounted to WNs

S.Chechelnitskiy / SFU Simon Fraser Test Results (continued) Test Scenario: One 4 cpu MPI job running on 2 different WNs, two local jobs and two Atlas jobs starting their XENs Play with priorities, queues Play with different images Test Results: Everything works, ready for implementation Preliminary results: migration of serial jobs (from node to node) also works Images are successfully stored/migrated/deleted

S.Chechelnitskiy / SFU Simon Fraser Test Results (continued) Advantages of this Configuration: Great flexibility in memory usage (Moab can handle various scenarios) Very efficient hardware usage Clusters can be configured/updated/upgraded almost independently Great flexibility: can migrate all serial jobs between nodes to free the whole WN(s) for MPI jobs

S.Chechelnitskiy / SFU Simon Fraser Implementation Schedule October Setup a small prototype of the future cluster: two final separate headnodes, 2 to 5 worker nodes November Test the prototype, optimize scheduling policies, verify the changes to LCG software November - December convert the prototype into a production 28+ nodes blade center cluster, run real WestGrid and Atlas jobs, replace the current SFU Atlas cluster March expand the small cluster into the final WestGrid/Atlas cluster (2000+ cores)

S.Chechelnitskiy / SFU Simon Fraser Running SE-dcache in XEN Motivation Reliability. In case of a hardware failure you can restart the virtual machine on another box in 5 minutes. Run XEN on openSolaris with ZFS and use ZFS advantages (improved mirroring, fast snapshots) to increase the stability and improve the emergency repair time.

S.Chechelnitskiy / SFU Simon Fraser Running SE-dcache in XEN (cont.) Limitations: Due to XEN’s poor networking properties only the core dCache services can run in a virtualized environment (no pools, no grdftp or dcap doors) Test Result: A base dCache-1.8 with Chimera namespace provider was installed in a XEN running on openSolaris-10.3 (guest OS was openSUSe- 10.2) Same dCache installations were made on the same hardware running openSolaris and openSUSe (to compare the performance) Pools and doors were started on different hardware

S.Chechelnitskiy / SFU Simon Fraser Running SE-dcache in XEN (cont.) Test Result: All functional tests were successful Performance was up to 95% of the same dCache configuration performance running at the same OS in a non- virtualized environment Benefit: run different dCache core services in different virtual machines and balance the load on the fly Non-XEN dCache runs approximately 10-15% faster on Solaris than on Linux (AMD box). Better java optimization, memory usage ?

S.Chechelnitskiy / SFU Simon Fraser Benefits of a Virtualized Environment Low Risk Scalability: Running jobs in a virtualized environment is possible and has a lot of advantages. Disadvantages are minor Lower Costs: Buy a general purpose cluster and use it for many different purposes Stability: Setup your site specific scheduling policies without affecting your Atlas cluster Flexibility: Run different LCG services in a virtualized environment to improve the stability and flexibility of your site

S.Chechelnitskiy / SFU Simon Fraser Acknowledgments! Thanks to Cluster Resources Inc. for their help in setting up the specific configuration And thanks also to SUN Microsystems Inc. for providing the hardware and software for our tests Questions?