ROCKS & The CASCI Cluster By Rick Bohn. What’s a Cluster? Cluster is a widely-used term meaning independent computers combined into a unified system through.

Slides:



Advertisements
Similar presentations
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
Advertisements

2. Computer Clusters for Scalable Parallel Computing
Beowulf Supercomputer System Lee, Jung won CS843.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Types of Parallel Computers
Information Technology Center Introduction to High Performance Computing at KFUPM.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
1 Week #1 Objectives Review clients, servers, and Windows network models Differentiate among the editions of Server 2008 Discuss the new Windows Server.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
© UC Regents 2010 Extending Rocks Clusters into Amazon EC2 Using Condor Philip Papadopoulos, Ph.D University of California, San Diego San Diego Supercomputer.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Grid Computing Net 535.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
11 SERVER CLUSTERING Chapter 6. Chapter 6: SERVER CLUSTERING2 OVERVIEW  List the types of server clusters.  Determine which type of cluster to use for.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Rocks Clusters SUN HPC Consortium November 2004 Federico D. Sacerdoti Advanced CyberInfrastructure Group San Diego Supercomputer Center.
Deploying and Managing Windows Server 2012
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
Chapter-4 Windows 2000 Professional Win2K Professional provides a very usable interface and was designed for use in the desktop PC. Microsoft server system.
Operational computing environment at EARS Jure Jerman Meteorological Office Environmental Agency of Slovenia (EARS)
December 8 & 9, 2005, Austin, TX SURA Cyberinfrastructure Workshop Series: Grid Technology: The Rough Guide Configuring Resources for the Grid Jerry Perez.
INSTALLING MICROSOFT EXCHANGE SERVER 2003 CLUSTERS AND FRONT-END AND BACK ‑ END SERVERS Chapter 4.
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
So, Jung-ki Distributed Computing System LAB School of Computer Science and Engineering Seoul National University Implementation of Package Management.
Rocks ‘n’ Rolls An Introduction to Programming Clusters using Rocks © 2008 UC Regents Anoop Rajendra.
CLUSTER COMPUTING STIMI K.O. ROLL NO:53 MCA B-5. INTRODUCTION  A computer cluster is a group of tightly coupled computers that work together closely.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
March 3rd, 2006 Chen Peng, Lilly System Biology1 Cluster and SGE.
Improving Network I/O Virtualization for Cloud Computing.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
SoCal Infrastructure OptIPuter Southern California Network Infrastructure Philip Papadopoulos OptIPuter Co-PI University of California, San Diego Program.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
Cluster Software Overview
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
(WINDOWS PLATFORM - ITI310 – S15)
1 Grid Activity Summary » Grid Testbed » CFD Application » Virtualization » Information Grid » Grid CA.
Virtualization Technology and Microsoft Virtual PC 2007 YOU ARE WELCOME By : Osama Tamimi.
Installation of Storage Foundation for Windows High Availability 5.1 SP2 1 Daniel Schnack Principle Technical Support Engineer.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Building and managing production bioclusters Chris Dagdigian BIOSILICO Vol2, No. 5 September 2004 Ankur Dhanik.
Cloud Computing – UNIT - II. VIRTUALIZATION Virtualization Hiding the reality The mantra of smart computing is to intelligently hide the reality Binary->
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Background Computer System Architectures Computer System Software.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
© 2007 UC Regents1 Rocks – Present and Future The State of Things Open Source Grids and Clusters Conference Philip Papadopoulos, Greg Bruno Mason Katz,
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Advanced Network Administration Computer Clusters.
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Clouds , Grids and Clusters
GWE Core Grid Wizard Enterprise (
Introduction to Operating System (OS)
Cloud based Open Source Backup/Restore Tool
University of Technology
Presentation transcript:

ROCKS & The CASCI Cluster By Rick Bohn

What’s a Cluster? Cluster is a widely-used term meaning independent computers combined into a unified system through software and networking. At the most fundamental level, when two or more computers are used together to solve a problem, it is considered a cluster.

Beowulf Cluster? Beowulf Clusters are scalable performance clusters based on commodity hardware, on a private system network, with open source software (Linux) infrastructure. The designer can improve performance proportionally with added machines. The commodity hardware can be any of a number of mass-market, stand-alone compute nodes as simple as two networked computers each running Linux and sharing a file system or as complex as 1024 nodes with a high-speed, low-latency network.

High Performance or High Throughput The key questions are: Granularity & Degree of Parallelism Have you got one big problem or a bunch of little ones? To what extent can the problem be decomposed into sort-of-independent parts (grains) that can all be processed in parallel? Granularity –Fine-grained parallelism – the independent bits are small, need to exchange information, synchronize often. –Coarse-grained – the problem can be decomposed into large chunks that can be processed independently.

HPC versus HTC Fine-grained problems need a high performance system –That enables rapid synchronization between the bits that can be processed in parallel –Runs the bits that are difficult to parallelize as fast as possible Coarse-grained problems can use a high throughput system –It maximizes the number of parts processed per minute HPC systems use a smaller number of more expensive processors expensively interconnected and is reliable HTC systems use a large number of inexpensive processors, inexpensively interconnected

Other Types of Clusters 1.Highly Available (HA) Generally small number of nodes Redundant components Multiple communication paths. 1.Visualization Clusters Each node drives a display OpenGL machines

Cluster Architecture Frontend Node Public Ethernet Private Ethernet Network Application Network (Optional) Node

So What’s a Grid? The term Grid computing originated in the early 1990s as a metaphor for making computer power as easy to access as an electric power grid.Today there are many definitions of Grid computing. IBM defines Grid Computing as "the ability, using a set of open standards and protocols, to gain access to applications and data, processing power, storage capacity and a vast array of other computing resources over the Internet. A Grid is a type of parallel and distributed system that enables the sharing, selection, and aggregation of resources distributed across 'multiple' administrative domains based on their (resources) availability, capacity, performance, cost and users' quality-of-service requirements" Grids can be categorized with a three stage model of departmental Grids, enterprise Grids and global Grids.

NYSGrid Status

Things to Consider Clusters are phenomenal price/performance computational engines. However They can be hard to manage without experience High-performance I/O is still evolving Finding out where something has failed increases at least linearly as cluster size increases Not cost-effective if every cluster “burns” a person just for care and feeding Programming environment could be vastly improved Technology is changing very rapidly. Scaling up is becoming commonplace

CASCI Cluster Center for Advancing the Study of Cyberinfrastructure (CASCI) Guy Johnson, Director

CASCI Cluster Hardware Head Node (1) IBM xSeries GB Ram 2 Pentium GHz 6 Hard Drives 36 GB (Internal RAID 5) 2 Gig Ethernet Ports Compute Nodes (47) IBM xSeries MB Ram 2 Pentium GHz 1 36 GB Hard Drive 1 Gig Ethernet Port

NYSGrid Cluster Hardware Head Node (1) IBM xSeries MB Ram 2 Pentium GHz 1 36 GB Hard Drive 2 Fast Ethernet Ports Compute Nodes (4) IBM xSeries MB Ram 2 Pentium GHz 1 36 GB Hard Drive 1 Fast Ethernet Port Experimental global grid cluster connected to other universities within New York state.

CASCI Cluster NETWORK The local network (eth0) is gigabit Ethernet using an Extreme Networks 6808 gigabit switch.

CASCI Cluster Images

The Great Wall of Cluster! Cluster courtesy of Paul Mezzanini Located behind CASCI Cluster racks

ROCKS Clustering Software

ROCKS Collaborators San Diego Supercomputer Center, UCSD Scalable Systems Pte Ltd in Singapore High Performance Computing Group, University of Tromso The Open Scalable Cluster Environment, Kasetsart University, Thailand Flow Physics and Computation Division, Stanford University Sun Microsystems Advanced Micro Devices

ROCKS Cluster Software Goal: Make Clusters Easy! 1.Easy to deploy, manage, upgrade and scale. 1.Help deliver the computational power of clusters to a wide range of scientific users. Making stable and manageable parallel computing platforms available to a wide range of scientists will aid immensely in improving the state of the art in parallel tools.

Supported Platforms ROCKS - is built on top of RedHat Linux releases (CentOS) - supports all the hardware components that RedHat supports, but only supports the x86, x86_64 and IA-64 architectures. Processors x86 (ia32, AMD Athlon, etc.) x86_64 (AMD Opteron and EM64T) IA-64 (Itanium) Networks Ethernet (All flavors that RedHat supports, including Intel Gigabit Ethernet) Myrinet (provided by Myricom) Infiniband (provided by Voltaire)

Minimum Hardware Requirements Frontend Node Disk Capacity: 20 GB Memory Capacity: 512 MB (i386) and 1 GB (x86_64) Ethernet: 2 physical ports (e.g., "eth0" and "eth1") Compute Node Disk Capacity: 20 GB Memory Capacity: 512 MB Ethernet: 1 physical port (e.g., "eth0")

ROCKS Distribution The ROCKS software is bundled into various packages called “Rolls” and put on CDs. Rolls are specially compiled to fit into the ROCKS installation methodology. Rolls are classified as either mandatory or optional. Rolls cannot be installed after the initial installation.

ROCKS Base Rolls The minimum requirements to bring up a frontend is to have the following Rolls. Kernel/Boot Roll Core Roll (Base, HPC, Web-server) OR BASE, HPC & Web-server Rolls Service Pack Roll OS Roll - Disk 1 OS Roll - Disk 2

ROCKS Optional Rolls The optional Rolls are: –Core Roll Area 51 (chkrootkit and tripwire) Ganglia (system monitoring software) Grid (software for connecting clusters) Java (Sun Java SDK and JVM) SGE (Sun Grid Engine scheduler) –Bio (bioinformatics utilities (release 4.2)) –Condor (high throughput computing tools) –PBS (portable batch scheduling software) –PVFS2 (parallel virtual file system version 2) –VIZ (visualization software) –Voltaire(Infiband support for Voltaire IB hardware)

ROCKS Software Stack

The Head Node Users login, submit jobs, compile code, etc Uses two Ethernet interfaces – one public, one private for compute nodes Normally has lots of disk space (system partitions < 14 GB) Provides many system services –NFS, DHCP, DNS, MySQL, HTTP, 411, Firewall,etc Cluster configuration

Compute Nodes Basic compute workhorse Lots of memory (if lucky) Minimal storage requirements Single Ethernet connection for private LAN Disposable OS easily re-installed from head node Nodes can be heterogeneous

NFS in ROCKS User accounts are served over NFS –Works for small clusters (< 128 nodes) –Will not work for large clusters (>1024) –NAS tends to work better Applications are not served over NFS –/usr/local does not exist –All software is installed locally (/opt)

411 Secure Information Service Provides NIS-like functionality Securely distributes password files, user and group configuration files and the like using Public Key Cryptography to protect file content. Uses HTTP to distribute the files Scalable, secure and low latency

411 Architecture 1.Client nodes listen on the IP broadcast address for “411 alert” messages from the head node. 2.Nodes then pull the file from the head node via HTTP after some delay to avoid flooding the master with requests.

As Simple as 411 To make changes to the 411 system you simply use “make” and the 411 “Makefile” similar to NIS. To publish 411 changes, on the head node run the command: 411put To retrieve 411 changes, on the compute node run the command: 411get or on the head node: cluster-fork 411get --all

Ganglia Monitoring Ganglia is a scalable distributed monitoring system for high- performance computing systems such as clusters and grids It leverages widely used technologies such as XML for data representation, XDR for compact, portable data transport, and RRDtool for data storage and visualization. It uses carefully engineered data structures and algorithms to achieve very low per-node overheads and high concurrency. Provides a heartbeat to determine compute node availability.

Cluster Status with Ganglia

Security Tools Tripwire runs everyday and s results. Chkrootkit is available and is manually executed Iptables is used as the firewall. Only trusted networks are allowed access.

Job Management Not recommended to run jobs directly! –Can hog cluster/nodes –No accountability Use installed job scheduler –You can submit multiple jobs and have it queued (and go home!) –Fair Share Allow other people to use the cluster also! - Accountability CASI Cluster Users! Without job management

Scheduling Systems Sun Grid Engine (default scheduler) –Rapidly becoming the new standard –Integrated into Rocks by Scalable System –Now the default scheduler for Rocks –Robust, dynamic and heterogeneous –Currently using 6.0 Portable Batch System(torque) and Maui –Long time standard for HPC queuing systems –Maui provides backfilling for high throughput –PBS/Maui system can be fragile and unstable –Multiple code bases: PBS, OpenPBS, etc Condor – high throughput computing ( currently under evaluation)

Sun Grid Engine (SGE) SGE is resource management software –Accepts jobs submitted by users –Schedules them for execution on appropriate systems based on resource management policies –Can submit 100s of jobs without worrying where it will run –Supports serial as well as parallel jobs

SUN Grid Engine Versions SGE Standard Edition –Linux cluster SGE Enterprise Edition –when you want to aggregate a few clusters together and manage them as one resource –When you want sophisticated policy management User/Project share Deadlines User, Department, Project level Rocks comes standard with SGE Enterprise 6.0

Cluster Web Site (

Requesting an Account

Accessing the Cluster Access the cluster via an SSH client PuTTY SSH Secure Shell X-Win32 F-Secure To transfer data to the cluster use either scp or sftp. Windows users can download and use WinSCP (

Available Applications BLAST (basic local alignment search tool for bio research) ENVI / IDL Data Visualization Software GCC (C, C++, Fortran programming) Mathematica (licensing limitations) Matlab (licensing limitations) mpiBLAST (parallel version of BLAST) MPICH (MPI parallel programming)

Other Alternatives to ROCKS Clustering Software Perceus / Warewulf ( openMosix Project (openmosix.sourceforge.net) Score Cluster System ( OSCAR (oscar.openclustergroup.org) System Imaging / Configuration Software System Imager (wiki.systemimager.org) Cfengine ( LCFG (

THANK YOU A Bad to the Bohn Production