Outline Course Administration Parallel Archtectures –Overview –Details Applications Special Approaches Our Class Computer Four Bad Parallel Algorithms.

Slides:



Advertisements
Similar presentations
Gordon Bell Bay Area Research Center Microsoft Corporation
Advertisements

1 Uniform memory access (UMA) Each processor has uniform access time to memory - also known as symmetric multiprocessors (SMPs) (example: SUN ES1000) Non-uniform.
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
© 2009 Fakultas Teknologi Informasi Universitas Budi Luhur Jl. Ciledug Raya Petukangan Utara Jakarta Selatan Website:
Prepared 7/28/2011 by T. O’Neil for 3460:677, Fall 2011, The University of Akron.
Distributed Systems CS
Clusters Part 1 - Definition of and motivation for clusters Lars Lundberg The slides in this presentation cover Part 1 (Chapters 1-4) in Pfister’s book.
Definitions of Supercomputer A time dependent term which refers to the class of most powerful computer systems world-wide at the time of reference An.
Zhao Lixing.  A supercomputer is a computer that is at the frontline of current processing capacity, particularly speed of calculation.  Supercomputers.
Parallel Programming Henri Bal Rob van Nieuwpoort Vrije Universiteit Amsterdam Faculty of Sciences.
Supercomputers Daniel Shin CS 147, Section 1 April 29, 2010.
Parallel Programming Henri Bal Vrije Universiteit Faculty of Sciences Amsterdam.
Multiprocessors ELEC 6200: Computer Architecture and Design Instructor : Agrawal Name: Nam.
Parallel Computers Past and Present Yenchi Lin Apr 17,2003.

1 Lecture 1: Parallel Architecture Intro Course organization:  ~5 lectures based on Culler-Singh textbook  ~5 lectures based on Larus-Rajwar textbook.
CPE 731 Advanced Computer Architecture Multiprocessor Introduction
Chapter 13 The First Component: Computer Systems.
1 Computer Science, University of Warwick Architecture Classifications A taxonomy of parallel architectures: in 1972, Flynn categorised HPC architectures.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Parallel Programming Henri Bal Vrije Universiteit Amsterdam Faculty of Sciences.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Parallel Architectures
1 In Summary Need more computing power Improve the operating speed of processors & other components constrained by the speed of light, thermodynamic laws,
CENG 546 Dr. Esma Yıldırım. Copyright © 2012, Elsevier Inc. All rights reserved What is a computing cluster?  A computing cluster consists of.
Computer System Architectures Computer System Software
1 Programming Multicore Processors Aamir Shafi High Performance Computing Lab
1 Lecture 20: Parallel and Distributed Systems n Classification of parallel/distributed architectures n SMPs n Distributed systems n Clusters.
Edgar Gabriel Short Course: Advanced programming with MPI Edgar Gabriel Spring 2007.
Parallel Processing - introduction  Traditionally, the computer has been viewed as a sequential machine. This view of the computer has never been entirely.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Message Passing Computing 1 iCSC2015,Helvi Hartmann, FIAS Message Passing Computing Lecture 1 High Performance Computing Helvi Hartmann FIAS Inverted CERN.
Copyright © 2011 Curt Hill MIMD Multiple Instructions Multiple Data.
Orange Coast College Business Division Computer Science Department CS 116- Computer Architecture Multiprocessors.
Parallel Computing.
Distributed Programming CA107 Topics in Computing Series Martin Crane Karl Podesta.
Outline Why this subject? What is High Performance Computing?
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
1 Lecture 1: Parallel Architecture Intro Course organization:  ~18 parallel architecture lectures (based on text)  ~10 (recent) paper presentations 
Understanding Parallel Computers Parallel Processing EE 613.
3/12/2013Computer Engg, IIT(BHU)1 INTRODUCTION-2.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 3.
CDA-5155 Computer Architecture Principles Fall 2000 Multiprocessor Architectures.
Background Computer System Architectures Computer System Software.
Primitive Concepts of Distributed Systems Chapter 1.
Hardware Trends CSE451 Andrew Whitaker. Motivation Hardware moves quickly OS code tends to stick around for a while “System building” extends way beyond.
COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques Dr. Xiao Qin Auburn University
1  2004 Morgan Kaufmann Publishers Fallacies and Pitfalls Fallacy: the rated mean time to failure of disks is 1,200,000 hours, so disks practically never.
Constructing a system with multiple computers or processors 1 ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson. Jan 13, 2016.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Introduction. News you can use Hardware –Multicore chips (2009: mostly 2 cores and 4 cores, but doubling) (cores=processors) –Servers (often.
These slides are based on the book:
Introduction to Parallel Processing
Flynn’s Taxonomy Many attempts have been made to come up with a way to categorize computer architectures. Flynn’s Taxonomy has been the most enduring of.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Introduction Super-computing Tuesday
Multiprocessor Systems
CS 147 – Parallel Processing
Constructing a system with multiple computers or processors
Grid Computing Colton Lewis.
Introduction.
Introduction.
Lecture 1: Parallel Architecture Intro
Different Architectures
Chapter 17 Parallel Processing
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Chapter 4 Multiprocessors
Vrije Universiteit Amsterdam
Presentation transcript:

Outline Course Administration Parallel Archtectures –Overview –Details Applications Special Approaches Our Class Computer Four Bad Parallel Algorithms

Parallel Computer Architectures MPP – Massively Parallel Processors –Top of the top500 list consists of mostly mpps but clusters are “rising” Clusters –“simple” cluster (1 processor in each node) –Cluster of small smp’s (small # processors / node) –Constellations (large # processors / node) Older Architectures –SIMD – Single Instruction Multiple Data (CM2) –Vector Processors (Old Cray machines)

Architecture Details 1.MPPs are built with speicalized networks by vendors with the intent of being used as a parallel computer. Clusters are built from independent computers integrated through an aftermarket network. Buzzwords: “COTS” Commodity off the shelf componets rather than custom archictures. Clusters are a market reactions to MPPS with the thought of being cheaper Originally considered to have solower communications but are catching up.

More details 2. “NOW”-Networks of Workstations Beowulf (Goddard in Greenbelt MD) – Clusters of a small number of pc’s (pre 10 th century poem in Old English about a Scandinaavian warrior from the 6 th century)

More Details Computers  SMPs P M World’s simplest computer P M C D Standard computer P M C D P M C D P M C D MPP

More Details SMP (Symmetric Multiprocessor) P/C M + Disk NUMA – Non uniform memory access 4. Constellation: Every nodes ia large smps

More details 5) SIMD – SIngle inmctruction multiple data 6: Speeds: Megaflops 10 6 flops Gigaflops 10 9 flops workstations Teraflops top 17 supercomputers by 2005 every supercomputer in the top 500 Petaflops ?

Moore’s Law: Number transistors per square in in an integrated circuit doubles every 18 months Every decade – computer performance increases 2 order of magnitude

Applications of Parallel Computers Traditionally: government labs, numerically intensive applications Research Institutions Recent Growth in Industrial Applications –236 of the top 500 –Financial analysis, drug design and analysis, oil exploration, aerospace and automotive

Goal of Parallel Computing Solve bigger problems faster Challenge of Parallel Computing Coordinate, control, and monitor the computation

Easiest Applictions Embarassingly Parallel – Lots of work that can be dived out with little coordination or communication Example: integration, Monte Carlo methods, Adding numbers

Special Approaches Distributed Computing on the internet signal processing 15 –Distributed.net factor product of 2 large primes –Parabon – biomedical, protein folding, gene expression Akamai Network – Tom Leighton, Danny Lewin Thousands of servers spread globally that caches web pages and routes traffic away from congested areas Embedded Computing : Mercury (inverse to the worldwide distribution)