Seaborg Cerise Wuthrich CMPS 5433. Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture.

Slides:



Advertisements
Similar presentations
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
Advertisements

Distributed Systems CS
Slides Prepared from the CI-Tutor Courses at NCSA By S. Masoud Sadjadi School of Computing and Information Sciences Florida.
Multiprocessors— Large vs. Small Scale Multiprocessors— Large vs. Small Scale.
WHAT IS AN OPERATING SYSTEM? An interface between users and hardware - an environment "architecture ” Allows convenient usage; hides the tedious stuff.
AMLAPI: Active Messages over Low-level Application Programming Interface Simon Yau, Tyson Condie,
Types of Parallel Computers
CSCI-455/522 Introduction to High Performance Computing Lecture 2.
Cache Coherent Distributed Shared Memory. Motivations Small processor count –SMP machines –Single shared memory with multiple processors interconnected.
Adam Kunk Anil John Pete Bohman.  Released by IBM in 2010 (~ February)  Successor of the POWER6  Implements IBM PowerPC architecture v2.06  Clock.
SHARCNET. Multicomputer Systems r A multicomputer system comprises of a number of independent machines linked by an interconnection network. r Each computer.
History of Distributed Systems Joseph Cordina
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
Reference: Message Passing Fundamentals.
CS 213 Commercial Multiprocessors. Origin2000 System – Shared Memory Directory state in same or separate DRAMs, accessed in parallel Upto 512 nodes (1024.
1 BGL Photo (system) BlueGene/L IBM Journal of Research and Development, Vol. 49, No. 2-3.
Multiprocessors ELEC 6200: Computer Architecture and Design Instructor : Agrawal Name: Nam.
1 Multiprocessors. 2 Idea: create powerful computers by connecting many smaller ones good news: works for timesharing (better than supercomputer) bad.
1: Operating Systems Overview
Server Platforms Week 11- Lecture 1. Server Market $ 46,100,000,000 ($ 46.1 Billion) Gartner.
OPERATING SYSTEM OVERVIEW
Hitachi SR8000 Supercomputer LAPPEENRANTA UNIVERSITY OF TECHNOLOGY Department of Information Technology Introduction to Parallel Computing Group.
IBM RS/6000 SP POWER3 SMP Jari Jokinen Pekka Laurila.
Adam Kunk Anil John Pete Bohman.  Released by IBM in 2010 (~ February)  Successor of the POWER6  Shift from high frequency to multi-core  Implements.
Fall 2008Introduction to Parallel Processing1 Introduction to Parallel Processing.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
1 Chapter 4 The Central Processing Unit and Memory.
Introduction to Symmetric Multiprocessors Süha TUNA Bilişim Enstitüsü UHeM Yaz Çalıştayı
Computer performance.
1 Lecture 7: Part 2: Message Passing Multicomputers (Distributed Memory Machines)
Computer System Architectures Computer System Software
LOGO OPERATING SYSTEM Dalia AL-Dabbagh
 What is an operating system? What is an operating system?  Where does the OS fit in? Where does the OS fit in?  Services provided by an OS Services.
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
1 Interconnects Shared address space and message passing computers can be constructed by connecting processors and memory unit using a variety of interconnection.
Multiprocessor systems Objective n the multiprocessors’ organization and implementation n the shared-memory in multiprocessor n static and dynamic connection.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Course Introduction Andy Wang COP 5611 Advanced Operating Systems.
Chapter 2 Parallel Architecture. Moore’s Law The number of transistors on a chip doubles every years. – Has been valid for over 40 years – Can’t.
High Performance Computing Processors Felix Noble Mirayma V. Rodriguez Agnes Velez Electric and Computer Engineer Department August 25, 2004.
1 Computer System Organization I/O systemProcessor Compiler Operating System (Windows 98) Application (Netscape) Digital Design Circuit Design Instruction.
April 2000Dr Milan Simic1 Network Operating Systems Windows NT.
Multiprocessing. Going Multi-core Helps Energy Efficiency William Holt, HOT Chips 2005 Adapted from UC Berkeley "The Beauty and Joy of Computing"
Frank Casilio Computer Engineering May 15, 1997 Multithreaded Processors.
Architectural Characterization of an IBM RS6000 S80 Server Running TPC-W Workloads Lei Yang & Shiliang Hu Computer Sciences Department, University of.
Architectural Characterization of an IBM RS6000 S80 Server Running TPC-W Workloads Lei Yang & Shiliang Hu Computer Sciences Department, University of.
BSP on the Origin2000 Lab for the course: Seminar in Scientific Computing with BSP Dr. Anne Weill –
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
Jacquard: Architecture and Application Performance Overview NERSC Users’ Group October 2005.
1 CMPE 511 HIGH PERFORMANCE COMPUTING CLUSTERS Dilek Demirel İşçi.
Computer Organization & Assembly Language © by DR. M. Amer.
Orange Coast College Business Division Computer Science Department CS 116- Computer Architecture Multiprocessors.
Interconnection network network interface and a case study.
Outline Why this subject? What is High Performance Computing?
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
3/12/2013Computer Engg, IIT(BHU)1 INTRODUCTION-2.
Spring EE 437 Lillevik 437s06-l22 University of Portland School of Engineering Advanced Computer Architecture Lecture 22 Distributed computer Interconnection.
CDA-5155 Computer Architecture Principles Fall 2000 Multiprocessor Architectures.
Background Computer System Architectures Computer System Software.
Chapter 11 System Performance Enhancement. Basic Operation of a Computer l Program is loaded into memory l Instruction is fetched from memory l Operands.
Intro to Distributed Systems Hank Levy. 23/20/2016 Distributed Systems Nearly all systems today are distributed in some way, e.g.: –they use –they.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Bus Systems ISA PCI AGP.
Andy Wang COP 5611 Advanced Operating Systems
CS775: Computer Architecture
Andy Wang COP 5611 Advanced Operating Systems
Chip&Core Architecture
Andy Wang COP 5611 Advanced Operating Systems
Cluster Computers.
Presentation transcript:

Seaborg Cerise Wuthrich CMPS 5433

Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture

Seaborg  Used by National Energy Research Scientific Computing Center (funded by Department of Energy) at Berkeley Lab  Named for Glenn Seaborg – Nobel laureate chemist who discovered 10 atomic elements, including plutonium

IBM SP RS/6000 Architecture  SP – Scalable Power Parallel  RS – RISC System  Composed of nodes

Nodes  416 nodes with 16 processors/node  380 compute nodes  20 nodes used for disk storage  6 login nodes  2 network nodes  8 spares

Front and back view of nodes

Node Architecture  16 IBM Power3 processors per node  Between 16 – 64 GB Memory per node  2 network adapters per node

Processors  IBM Power3 processors each running at 375 MHz  Power – Performance Optimized With Enhanced RISC  PowerPC processors are RISC-based symmetric multiprocessors (every processor is functionally identical) with 64-bit addressability  Connected to L2 cache by bus running at 250 MHz  Dynamic Branch Prediction  Instruction prefetching  FP units are fully pipelined  4 FLOP/cycle x 375 MHz = 1500 Million or 1.5 GFLOPS/sec

Power PC 3 processor 32 KB 64KB 8 MB

Power3 Processor 15 million transistors

Interconnection Network  Nodes connected with high bandwidth, low latency IBM SP2 switch  Can be connected in various topologies depending on number of nodes  Each switchboard has up to 32 links  16 links to nodes  16 links to other switchboards

Interconnection Network Star Topology used for up to 80 nodes and still guarantee 4 independent shortest paths

Interconnection Network Intermediate switchboards must be added for nodes

Interconnection Network  The combination of HW and SW of the switch system is known as the CSS – Communication SubSystem  Network is highly available

Latency in the network  Within nodes, latency is 9 microseconds  Between nodes, using Message Passing Interface, the latency is 17 microseconds

Scalability  Architecture can handle from 1 – 512 nodes  The current version of Seaborg (2003) is twice the size of the original (2001)

Memory shared memory  Within each node, between 16 & 64 GB of shared memory  Between nodes, there is distributed memory  Parallel programs can be run using distributed memory message passing, shared memory threading or a combination

I/O  20 nodes run the distributed parallel I/O system called GPFS – General Parallel File System  44 Terabytes of disk space  Each node runs its own copy of AIX – IBM’s Unix-based OS

Production Status/Cost  $33 Million for the first version put into operation in June 2001  At the time, it was the 2 nd most powerful computer in the world and the most powerful one for unclassified research  In 2003, number of nodes was doubled

Customers  2100 researchers at national labs and universities across the country  Restricted to Department of Energy funded massively parallel processing projects  Located at the National Energy Research Computing Center

Applications  Massively Parallel Scientific Research  Gasoline Combustion Simulation  Fusion Energy Research  Climate Modeling  Materials Science  Computational Biology  Particle Simulations  Plasma Acceleration  Large Scale Simulation of Atomic Structures

Interesting Features  In 2004, 2.4 times as many requests as resources available  Uses POE (Parallel Operating Environment) and LoadLeveler to schedule jobs

Survey Results – Why do you use Seaborg?  Need massively parallel computer  High speed  Achieves level of numerical accuracy  Can run several simulations in parallel  Easy to connect using ssh  Fastest and most efficient computer available for my research  Long queue times are great  Large enough memory for my needs

Survey Results – How could Seaborg be improved?  I think too many nodes are scheduled for many jobs. Scaling is not good in many cases.  “..Virtually impossible to do interactive work”  “Debuggers are terrible.”  “Compilers and debuggers are a step down from the Cray.”  Giving preference to high concurrency jobs makes smaller jobs wait

Questions