Copyright Gordon Bell Clusters & Grids The CC – GRID? Era Infinite processing, storage, and zero cost and latency Gordon Bell

Slides:



Advertisements
Similar presentations
Gordon Bell Bay Area Research Center Microsoft Corporation
Advertisements

Chapter 2 Introduction to Computer Networks INTRODUCTION TO COMPUTER NETWORKS.
Distributed Data Processing
2. Computer Clusters for Scalable Parallel Computing
Beowulf Supercomputer System Lee, Jung won CS843.
Crays, Clusters, Centers and Grids
Copyright Gordon Bell LANL 5/17/2002 Technical computing: Observations on an ever changing, occasionally repetitious, environment Los Alamos National Laboratory.
NPACI Panel on Clusters David E. Culler Computer Science Division University of California, Berkeley
Copyright Gordon Bell & Jim Gray PC+ The PC+ Era Infinite processing, memory, and zero cost. Gordon Bell Bay Area Research Center Microsoft.
MASPLAS ’02 Creating A Virtual Computing Facility Ravi Patchigolla Chris Clarke Lu Marino 8th Annual Mid-Atlantic Student Workshop On Programming Languages.
Revolution Yet to Happen1 The Revolution Yet to Happen Gordon Bell & James N. Gray (from Beyond Calculation, Chapter 1) Rivier College, CS699 Professional.
CPE 731 Advanced Computer Architecture Multiprocessor Introduction
Cluster Computing Slides by: Kale Law. Cluster Computing Definition Uses Advantages Design Types of Clusters Connection Types Physical Cluster Interconnects.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Hardware and Software Basics. Computer Hardware  Central Processing Unit - also called “The Chip”, a CPU, a processor, or a microprocessor  Memory (RAM)
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
F1031 COMPUTER HARDWARE CLASSES OF COMPUTER. Classes of computer Mainframe Minicomputer Microcomputer Portable is a high-performance computer used for.
System Architecture & Hardware Configurations Dr. D. Bilal IS 592 Spring 2005.
Edge Based Cloud Computing as a Feasible Network Paradigm(1/27) Edge-Based Cloud Computing as a Feasible Network Paradigm Joe Elizondo and Sam Palmer.
Hardware -Computer Organization. Hardware & Software A computer system consists of A computer system consists of –Hardware: anything you can touch, smell,
Types of Operating System
Information Systems Today: Managing in the Digital World TB4-1 4 Technology Briefing Networking.
1 Computing platform Andrew A. Chien Mohsen Saneei University of Tehran.
1 In Summary Need more computing power Improve the operating speed of processors & other components constrained by the speed of light, thermodynamic laws,
Introduction to Computers Personal Computing 10. What is a computer? Electronic device Performs instructions in a program Performs four functions –Accepts.
Chapter 5 Networks Communicating and Sharing Resources
Flash Cards Computer Technology.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
CLUSTER COMPUTING STIMI K.O. ROLL NO:53 MCA B-5. INTRODUCTION  A computer cluster is a group of tightly coupled computers that work together closely.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
March 22, 2000Dr. Thomas Sterling, Caltech1. Networking Options for Beowulf Clusters Dr. Thomas Sterling California Institute of Technology and NASA Jet.
Introduction to information technology Dr.Fatimah Ali Al-Rowibah.
Types of Operating Systems
C5- IT Infrastructure and Emerging Technologies. Input – Process - Output 2 A computer  Takes data as input  Processes it  Outputs information CPU.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Chapter 1 Communication Networks and Services Network Architecture and Services.
Large Scale Parallel File System and Cluster Management ICT, CAS.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
CLUSTER COMPUTING TECHNOLOGY BY-1.SACHIN YADAV 2.MADHAV SHINDE SECTION-3.
20 October Management of Information Technology Chapter 6 Chapter 6 IT Infrastructure and Platforms Asst. Prof. Wichai Bunchua.
Computing Environment The computing environment rapidly evolving ‑ you need to know not only the methods, but also How and when to apply them, Which computers.
IDE disk servers at CERN Helge Meinhard / CERN-IT CERN OpenLab workshop 17 March 2003.
Types of Operating Systems 1 Computer Engineering Department Distributed Systems Course Assoc. Prof. Dr. Ahmet Sayar Kocaeli University - Fall 2015.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture CSE 8383 April 11, 2006 Session 23.
Distributed Computing Systems CSCI 6900/4900. Review Distributed system –A collection of independent computers that appears to its users as a single coherent.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Copyright Gordon Bell Clusters & Grids The CC – GRID? Era CC GS C 2002 Gordon Bell Bay Area Research Center Microsoft Corporation.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
Tackling I/O Issues 1 David Race 16 March 2010.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Mbps over 5,626 km ~ 4e15 bit meters per second 4 Peta Bmps (“peta bumps”) Single Stream TCP/IP throughput Information Sciences Institute Microsoft.
Background Computer System Architectures Computer System Software.
Intro to Distributed Systems Hank Levy. 23/20/2016 Distributed Systems Nearly all systems today are distributed in some way, e.g.: –they use –they.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques Dr. Xiao Qin Auburn University
COMP7500 Advanced Operating Systems
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Berkeley Cluster Projects
Parallel Computers Definition: “A parallel computer is a collection of processing elements that cooperate and communicate to solve large problems fast.”
Overview Introduction to Operating Systems
What is Parallel and Distributed computing?
Ch 4. The Evolution of Analytic Scalability
Chapter 4 Multiprocessors
Cluster Computers.
Presentation transcript:

Copyright Gordon Bell Clusters & Grids The CC – GRID? Era Infinite processing, storage, and zero cost and latency Gordon Bell Bay Area Research Center Microsoft Corporation

Copyright Gordon Bell Clusters & Grids

deja’ vu ARPAnet: c1969 – To use remote programs & data – Got FTP & mail. Machines & people overloaded. NREN: c1988 – BW => Faster FTP for images, data – Latency => Got – Tomorrow => Gbit communication BW, latency <’90 Mainframes, minis, PCs/WSs >’90 very large, dep’t, & personal clusters VAX: c1979 one computer/scientist Beowulf: c1995 one computer/scientist 1960s batch: opti-use allocate, schedule,$ 2000s GRID: opti-use allocate, schedule, $ (… security, management, etc.)

Copyright Gordon Bell Clusters & Grids Some observations Clusters are purchased, managed, and used as a single, one room facility. Clusters are the “new” computers. They present unique, interesting, and critical problems… then Grids can exploit them. Clusters & Grids have little to do with one another… Grids use clusters! Clusters should be a good simulation of tomorrow’s Grid. Distributed PCs: Grids or Clusters? Perhaps some clusterable problems can be solved on a Grid… but it’s unlikely. – Lack of understanding clusters & variants – Socio-, political, eco- wrt to Grid.

Copyright Gordon Bell Clusters & Grids Some observations GRID was/is an exciting concept … – They can/must work within a community, organization, or project. What binds it? – “Necessity is the mother of invention.” Taxonomy… interesting vs necessity – Cycle scavenging and object evaluation (e.g. – File distribution/sharing aka IP theft (e.g. Napster, Gnutella) – Databases &/or programs and experiments (astronomy, genome, NCAR, CERN) – Workbenches: web workflow chem, bio… – Single, large problem pipeline… e.g. NASA. – Exchanges… many sites operating together – Transparent web access aka load balancing – Facilities managed PCs operating as cluster!

Copyright Gordon Bell Clusters & Grids Grids: Why? The problem or community dictates a Grid Economics… thief or scavenger Research funding… that’s where the problems are

In a 5-10 years we can/will have: more powerful personal computers – processing x; multiprocessors-on-a-chip – 4x resolution (2K x 2K) displays to impact paper – Large, wall-sized and watch-sized displays – low cost, storage of one terabyte for personal use adequate networking? PCs now operate at 1 Gbps – ubiquitous access = today’s fast LANs – Competitive wireless networking One chip, networked platforms e.g. light bulbs, cameras Some well-defined platforms that compete with the PC for mind (time) and market share watch, pocket, body implant, home (media, set-top) Inevitable, continued cyberization… the challenge… interfacing platforms and people.

Copyright Gordon Bell Clusters & Grids SNAP … c1995 S calable N etwork A nd P latforms A View of Computing in We all missed the impact of WWW! Gordon Bell Jim Gray

How Will Future Computers Be Built? Thesis: SNAP: Scalable Networks and Platforms Upsize from desktop to world-scale computer based on a few standard components Because: Moore’s law: exponential progress Standardization & Commoditization Stratification and competition When: Sooner than you think! Massive standardization gives massive use Economic forces are enormous

Copyright Gordon Bell Clusters & Grids Performance versus time for various microprocessors

Copyright Gordon Bell Clusters & Grids Volume drives simple, cost to standard platforms performanceperformance Stand-alone Desk tops PCs

Copyright Gordon Bell Clusters & Grids Computing SNAP built entirely from PCs Wide & Local Area Networks for: terminal, PC, workstation, & servers Centralized & departmental uni- & mP servers (UNIX & NT) Legacy mainframes & minicomputers servers & terms Wide-area global network Legacy mainframe & minicomputer servers & terminals Centralized & departmental servers buit from PCs scalable computers built from PCs TC=TV+PC home... (CATV or ATM or satellite) ??? Portables A space, time (bandwidth), & generation scalable environment Person servers (PCs) Person servers (PCs) Mobile Nets

Copyright Gordon Bell Clusters & Grids SNAP Architecture

Copyright Gordon Bell Clusters & Grids GB plumbing from the baroque: evolving from the 2 dance-hall model Mp — S — Pc : | : |——————-- S.fiber ch. — Ms | : |— S.Cluster |— S.WAN — vs, MpPcMs — S.Lan/Cluster/Wan — :

Copyright Gordon Bell Clusters & Grids Modern scalable switches … also hide a supercomputer Scale from <1 to 120 Tbps 1 Gbps ethernet switches scale to 10s of Gbps, scaling upward SP2 scales from 1.2

Copyright Gordon Bell Clusters & Grids Interesting “cluster” in a cabinet 366 servers per 44U cabinet – Single processor – GB/computer (24 TBytes) – Mbps Ethernets ~10x perf*, power, disk, I/O per cabinet ~3x price/perf Network services… Linux based *42, 2 processors, 84 Ethernet, 3 TBytes

Copyright Gordon Bell Clusters & Grids High Performance Computing A 60+ year view

Copyright Gordon Bell Clusters & Grids High performance architecture/program timeline VtubesTrans.MSI(mini) Micro RISCnMicr Sequential programming----> (single execution stream) <SIMD Vector--// Parallelization--- Parallel programs aka Cluster Computing < multicomputers <--MPP era ultracomputers 10X in size & price!10x MPP “in situ” resources 100x in //sm NOW CC geographically dispersedGrid

Copyright Gordon Bell Clusters & Grids Computer types Netwrked Supers… Legion Condor Beowulf NT clusters VPPuni T3E SP2 (mP) NOW NEC mP SGI DSM clusters & SGI DSM NEC super Cray X…T (all mPv) Mainframes Multis WSs PCs Connectivity WAN/LAN SAN DSM SM micros vector Clusters GRID Old World

Copyright Gordon Bell Clusters & Grids ISTORE Hardware Vision System-on-a-chip enables computer, memory, without significantly increasing size of disk 5-7 year target: MicroDrive:1.7” x 1.4” x 0.2” 2006: ? 1999: 340 MB, 5400 RPM, 5 MB/s, 15 ms seek 2006: 9 GB, 50 MB/s ? (1.6X/yr capacity, 1.4X/yr BW) Integrated IRAM processor 2x height Connected via crossbar switch growing like Moore’s law 16 Mbytes; ; 1.6 Gflops; 6.4 Gops 10,000+ nodes in one rack! 100/board = 1 TB; 0.16 Tf

Copyright Gordon Bell Clusters & Grids The Disk Farm? or a System On a Card? The 500GB disc card An array of discs Can be used as 100 discs 1 striped disc 50 FT discs....etc LOTS of accesses/second of bandwidth A few disks are replaced by 10s of Gbytes of RAM and a processor to run Apps!! 14"

Copyright Gordon Bell Clusters & Grids The Network Revolution Networking folks are finally streamlining LAN case (SAN). Offloading protocol to NIC ½ power point is 8KB Min round trip latency is ~50 µs. 3k ins +.1 ins/byte High-Performance Distributed Objects over a System Area Network Li, L. ; Forin, A. ; Hunt, G. ; Wang, Y., MSR-TR-98-68

Telnet & FTP WWW AudioVideo Voice! Standards Increase Capacity (circuits & bw) Lower response time Create new service Increased Demand The virtuous cycle of bandwidth supply and demand

Copyright Gordon Bell Clusters & Grids 744Mbps over 5000 km to transmit 14 GB ~ 4e15 bit meters per second 4 Peta Bmps (“peta bumps”) Single Stream tcp/ip throughput Information Sciences Institute Microsoft QWest University of Washington Pacific Northwest Gigapop HSCC (high speed connectivity consortium) DARPA

Copyright Gordon Bell Clusters & Grids Map of Gray Bell Prize results Redmond/Seattle, WA San Francisco, CA New York Arlington, VA 5626 km 10 hops

Copyright Gordon Bell Clusters & Grids 1 GBps Ubiquitous 10 GBps SANs in 5 years 1Gbps Ethernet are reality now. – Also FiberChannel,MyriNet, GigaNet, ServerNet,, ATM,… 10 Gbps x4 WDM deployed now (OC192) – 3 Tbps WDM working in lab In 5 years, expect 10x !! 5 MBps 20 MBps 40 MBps 80 MBps 120 MBps (1Gbps)

Copyright Gordon Bell Clusters & Grids The Promise of SAN/VIA:10x in 2 years Yesterday: – 10 MBps (100 Mbps Ethernet) – ~20 MBps tcp/ip saturates 2 cpus – round-trip latency ~250 µs Now – Wires are 10x faster Myrinet, Gbps Ethernet, ServerNet,… – Fast user-level communication - tcp/ip ~ 100 MBps 10% cpu - round-trip latency is 15 us 1.6 Gbps demoed on a WAN

Copyright Gordon Bell Clusters & Grids Labscape 1 st, 2 nd, 3 rd, or New Paradigm for science?

Copyright Gordon Bell Clusters & Grids Labscape

Copyright Gordon Bell Clusters & Grids Labscape

Copyright Gordon Bell Clusters & Grids Labscape sensors Location tracking of people/samples – multiple resolutions – passive and active tags Manual tasks (e.g., use of reagents, tools) Audio/video records, vision and indexing Networked instruments (e.g., pipettes, refrigerators, etc.)

Courtesy of Dr. Thomas Sterling, Caltech

Lessons from Beowulf An experiment in parallel computing systems Established vision- low cost high end computing Demonstrated effectiveness of PC clusters for some (not all) classes of applications Provided networking software Provided cluster management tools Conveyed findings to broad community Tutorials and the book Provided design standard to rally community!* Standards beget: books, trained people, software … virtuous cycle* *observations Courtesy, Thomas Sterling, Caltech.

Copyright Gordon Bell Clusters & Grids The End How can GRIDs become a non- ad hoc computer structure? Get yourself an application community!