Managing Linux Clusters with Rocks Tim Carlson - PNNL

Slides:



Advertisements
Similar presentations
QMUL e-Science Research Cluster Introduction (New) Hardware Performance Software Infrastucture What still needs to be done.
Advertisements

S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment S.Chechelnitskiy Simon Fraser University CHEP 2007 September 6 th.
Beowulf Supercomputer System Lee, Jung won CS843.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Overview of Midrange Computing Resources at LBNL Gary Jung March 26, 2002.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Job Submission on WestGrid Feb on Access Grid.
CSC Site Update HP Nordic TIG April 2008 Janne Ignatius Marko Myllynen Dan Still.
Academic and Research Technology (A&RT)
Cloud Don McGregor Research Associate MOVES Institute
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Capacity Planning in SharePoint Capacity Planning Process of evaluating a technology … Deciding … Hardware … Variety of Ways Different Services.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Building a High-performance Computing Cluster Using FreeBSD BSDCon '03 September 10, 2003 Brooks Davis, Michael AuYeung, Gary Green, Craig Lee The Aerospace.
Installing and maintaining clusters of FreeBSD servers using PXE and Rsync Cor Bosman XS4ALL
Rocks Clusters SUN HPC Consortium November 2004 Federico D. Sacerdoti Advanced CyberInfrastructure Group San Diego Supercomputer Center.
1 The Virtual Reality Virtualization both inside and outside of the cloud Mike Furgal Director – Managed Database Services BravePoint.
Rocks cluster : a cluster oriented linux distribution or how to install a computer cluster in a day.
Parallel Computing The Bad News –Hardware is not getting faster fast enough –Too many architectures –Existing architectures are too specific –Programs.
Computer System Architectures Computer System Software
A Makeshift HPC (Test) Cluster Hardware Selection Our goal was low-cost cycles in a configuration that can be easily expanded using heterogeneous processors.
Nagios Is Down and Your Boss Wants to See You Andrew Widdersheim
Introduction to HPC resources for BCB 660 Nirav Merchant
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
So, Jung-ki Distributed Computing System LAB School of Computer Science and Engineering Seoul National University Implementation of Package Management.
University of Southampton Clusters: Changing the Face of Campus Computing Kenji Takeda School of Engineering Sciences Ian Hardy Oz Parchment Southampton.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
October, Scientific Linux INFN/Trieste B.Gobbo – Compass R.Gomezel - T.Macorini - L.Strizzolo INFN - Trieste.
Rocks ‘n’ Rolls An Introduction to Programming Clusters using Rocks © 2008 UC Regents Anoop Rajendra.
การติดตั้งและทดสอบการทำคลัสเต อร์เสมือนบน Xen, ROCKS, และไท ยกริด Roll Implementation of Virtualization Clusters based on Xen, ROCKS, and ThaiGrid Roll.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Common Practices for Managing Small HPC Clusters Supercomputing 12
CMAQ Runtime Performance as Affected by Number of Processors and NFS Writes Patricia A. Bresnahan, a * Ahmed Ibrahim b, Jesse Bash a and David Miller a.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Grid MP at ISIS Tom Griffin, ISIS Facility. Introduction About ISIS Why Grid MP? About Grid MP Examples The future.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Computer Systems Lab The University of Wisconsin - Madison Department of Computer Sciences Linux Clusters David Thompson
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
Deploying a Network of GNU/Linux Clusters with Rocks / Arto Teräs Slide 1(18) Deploying a Network of GNU/Linux Clusters with Rocks Arto Teräs.
Status of Florida Tier2 Center A mini tutorial on ROCKS appliances Jorge L. Rodriguez February 2003.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Cluster Software Overview
U N I V E R S I T Y O F S O U T H F L O R I D A Hadoop Alternative The Hadoop Alternative Larry Moore 1, Zach Fadika 2, Dr. Madhusudhan Govindaraju 2 1.
RAL Site report John Gordon ITD October 1999
HEP Computing Status Sheffield University Matt Robinson Paul Hodgson Andrew Beresford.
1 Copyright © 2015 Pexus LLC Patriot PS Personal Server Installing Patriot PS ISO Image on.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
Version Control and SVN ECE 297. Why Do We Need Version Control?
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Tackling I/O Issues 1 David Race 16 March 2010.
Operational and Application Experiences with the Infiniband Environment Sharon Brunett Caltech May 1, 2007.
2: Operating Systems Networking for Home & Small Business.
Background Computer System Architectures Computer System Software.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarksEGEE-III INFSO-RI MPI on the grid:
© 2007 UC Regents1 Rocks – Present and Future The State of Things Open Source Grids and Clusters Conference Philip Papadopoulos, Greg Bruno Mason Katz,
CFI 2004 UW A quick overview with lots of time for Q&A and exploration.
1 Policy Based Systems Management with Puppet Sean Dague
Happy Endings: Reengineering Wesleyan’s Software Deployment to Labs and Classrooms Kyle Tousignant 03/22/2016.
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Super Computing By RIsaj t r S3 ece, roll 50.
Is System X for Me? Cal Ribbens Computer Science Department
Presentation transcript:

Managing Linux Clusters with Rocks Tim Carlson - PNNL

Introduction Cluster Design The ins and outs of designing compute solutions for scientists Rocks Cluster Software What it is and some basic philosophies of Rocks Midrange computing with Rocks at PNNL How PNNL uses Rocks to manage 25 clusters ranging from 32 to 1500 compute cores

I Need a Cluster! Can you make use of existing resources? chinook 2310 Barcelona CPUs with DDR Infiniband Requires EMSL proposal superdome 256 core Itanium 2 SMP machine Short proposal required Department clusters HPCaNS manages 25 clusters. Does your department have one of them? Limited amount of PNNL “general purpose” compute cycles

I Really Need a Cluster! Why? Run bigger models? Maybe you need a large memory desk side machine. 72G in a desk side is doable (dual Nehalem with 18 x 4G DIMMS) Do you need/want to run parallel code? Again, maybe a desk side machine is appropriate. 8 cores in single machine

You Need a Cluster What software do you plan to run? WRF/MM5 (atmospheric/climate) May benefit from low latency network Quad core scaling? NWChem (molecular chemistry) Usually requires a low latency network Need an interconnec that is fully supported by ARMCI/GA Fast local scratch required. Fast global scratch a good idea Home Grown Any idea of the profile of your code? Can we have a test case to run on our test cluster?

Processor choices Intel Harpertown or Nehalem Do you need the Nehalem memory bandwidth? AMD Barcelona or Shanghai Shanghai is a better Barcelona Disclaimer This talk was due 4 weeks early. All of the above could have changed in that time

More Hardware Choices Memory per core Be careful configuring Nehalem Interconnect GigE, DDR, QDR Local disk I/O Do you even use this? Global file system At any reasonable scale you probably aren’t using NFS Lustre/PVFS2/Panasas

Rocks Software Stack Redhat based PNNL is mostly Redhat so the environment is familiar NFS Funded since 2000 Several HPC Wire awards Our choice since 2001 Originally based on Redhat 6.2, now based on RHEL 5.3

Rocks is a Cluster Framework Customizable Not locked into a vendor solution Modify default disk partitioning Use your own custom kernel Add software via RPMs or “Rolls” Need to make more changes? Update an XML file, rebuild the distribution, reinstall all the nodes Rocks is not “system imager” based All nodes are “installed” and not “imaged”

Rocks Philosophies Quick to install It should not take a month (or even more than a day) to install a thousand node cluster Nodes are 100% configured No “after the fact” tweaking If a node is out of configuration, just reinstall Don’t spend time on configuration management of nodes Just reinstall

What is a Roll A Roll is a collection of software packages and configuration information “Rolls” provide more specific tools Commercial compiler Rolls (Intel, Absoft, Portland Group) Your choice of scheduler (Sun Grid Engine, Torque) Science specific (Bio Roll) Many others (Java, Xen, PVFS2, TotalView, etc) Users can build their own Rolls –

Scalable Not “system imager” based Non-homogeneous makes “system imager” types installation problematic Nodes install from kickstart files generated from a database Several clusters registered with over 500 nodes Avalanche installer removes pressure from any single installation server Introduced in Rocks 4.1 Torrent based Nodes share packages during installation

Community and Commercial Support Active mailing list averaging over 700 posts per month Annual “Rocks-A-Palooza” meeting for community members Talks, tutorials, working groups Rocks cluster register has over 1100 clusters registered representing more than 720 Teraflops of computational power ClusterCorp sells Rocks+ support based on open source Rocks

PNNL Midrange Clusters Started in node VALinux cluster Dual PIII 500Mhz with 10/100 ethernet Chose “Rocks” as the software stack Built our first “big” cluster that same year 64 Dual Pentium III at 1 Ghz Rebuild all the nodes with Rocks in under 30 minutes Parts of this system are still in production Currently manage 25 clusters Range in size from 16 to1536 cores Infiniband is the primary interconnect Attached storage ranges from 1 to 100 Terabytes 14

HPCaNS Management Philosophy Create service center to handle money Charge customers between $300 and $800/month based on size and complexity Covers account management, patching, minimal backups (100G), compiler licenses, BigBrother monitoring, general sysadmin Use.75 FTE to manage all the clusters “Non-standard” needs are charged by time and materials Adding new nodes Rebuilding to a new OS Software porting or debugging Complex queue configurations

Support Methods BigBrother alerts Hooks into ganglia checking for Node outages Disk usage problems to cluster sysadmins See next slide after a bad power outage! Support queue Users pointed to central support queue 5 UNIX admins watching the queue for cluster items Try to teach users to use the support queue

Typical Daily Questions Can you add application X, Y, Z? My job doesn’t seem to be running in the queue? The compiler gives me this strange error! Do you have space/power/cooling for this new cluster I want to buy? This code runs on cluster X, but doesn’t run on cluster Y. Why is that? Aren’t they the same? Can I add another 10T of disk storage? The cluster is broken!

Always Room for Improvement Clusters live in 4 different computer rooms Can we consolidate? Never enough user documentation Standardize on resource managers Currently have various versions of Torque and SLURM Should we be upgrading older OSes ? Still have RHEL 3 based clusters Do we need to be doing “shared/grid/cloud” computing? Why in the world do you have 25 clusters?

Questions, comments, discussion!