UCSD SAN DIEGO SUPERCOMPUTER CENTER 1 Who needs a supercomputer? Professor Snavely, University of California Professor Allan Snavely University of California,

Slides:



Advertisements
Similar presentations
ISO 9126: Software Quality Internal Quality: Maintainability
Advertisements

Hardware specification. The “brain” The brain processes instructions, performs calculations, and manages the flow of information through the body The.
1 Computational models of the physical world Cortical bone Trabecular bone.
FUTURE TECHNOLOGIES Lecture 13.  In this lecture we will discuss some of the important technologies of the future  Autonomic Computing  Cloud Computing.
2. Computer Clusters for Scalable Parallel Computing
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
Parallel Programming Henri Bal Rob van Nieuwpoort Vrije Universiteit Amsterdam Faculty of Sciences.
CS CS 5150 Software Engineering Lecture 19 Performance.
1 History of Computers Module 1 Section 1 Source: nfo.lindows.com/examples/powerpoint_example2.ppt.
MD240 - Management Information Systems Sept. 13, 2005 Computing Hardware – Moore's Law, Hardware Markets, and Computing Evolution.
1 COMP 206: Computer Architecture and Implementation Montek Singh Mon., Sep 5, 2005 Lecture 2.
What is Grid Computing? Grid Computing is applying the resources of many computers in a network to a single entity at the same time;  Usually to a scientific.
CS 501: Software Engineering Fall 2000 Lecture 19 Performance of Computer Systems.
Introduction What is Parallel Algorithms? Why Parallel Algorithms? Evolution and Convergence of Parallel Algorithms Fundamental Design Issues.
1 Computer Science, University of Warwick Metrics  FLOPS (FLoating point Operations Per Sec) - a measure of the numerical processing of a CPU which can.
CS CS 5150 Software Engineering Lecture 25 Performance.
1 CS 501 Spring 2005 CS 501: Software Engineering Lecture 22 Performance of Computer Systems.
Parallel Algorithms - Introduction Advanced Algorithms & Data Structures Lecture Theme 11 Prof. Dr. Th. Ottmann Summer Semester 2006.
Lecture 1: Introduction to High Performance Computing.
Energy Model for Multiprocess Applications Texas Tech University.
M206 – Data Measurement. Introduction ‘Have you ever wondered how the computer interprets data?’ This is the language that the computer understands. This.
CPU Performance Assessment As-Bahiya Abu-Samra *Moore’s Law *Clock Speed *Instruction Execution Rate - MIPS - MFLOPS *SPEC Speed Metric *Amdahl’s.
1.Training and education 2.Consulting 3.Travel 4.Hardware 5.Software Which of the following is not included in a firm’s IT infrastructure investments?
Department of Computer and Information Science, School of Science, IUPUI Dale Roberts, Lecturer Computer Science, IUPUI CSCI.
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
1 CHAPTER 2 COMPUTER HARDWARE. 2 The Significance of Hardware  Pace of hardware development is extremely fast. Keeping up requires a basic understanding.
Cell Architecture. Introduction The Cell concept was originally thought up by Sony Computer Entertainment inc. of Japan, for the PlayStation 3 The architecture.
High-Performance Computing 12.1: Concurrent Processing.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture.
 Design model for a computer  Named after John von Neuman  Instructions that tell the computer what to do are stored in memory  Stored program Memory.
Lecture 03: Fundamentals of Computer Design - Trends and Performance Kai Bu
Did You Know?. Sometimes size does matter? If you’re one in a million in China...
Lecture 1: Performance EEN 312: Processors: Hardware, Software, and Interfacing Department of Electrical and Computer Engineering Spring 2013, Dr. Rozier.
The Central Processing Unit
1 CS 501 Spring 2006 CS 501: Software Engineering Lecture 22 Performance of Computer Systems.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 February Session 6.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
Parallel Processing Steve Terpe CS 147. Overview What is Parallel Processing What is Parallel Processing Parallel Processing in Nature Parallel Processing.
Parallel Processing Sharing the load. Inside a Processor Chip in Package Circuits Primarily Crystalline Silicon 1 mm – 25 mm on a side 100 million to.
Thermal-aware Issues in Computers IMPACT Lab. Part A Overview of Thermal-related Technologies.
Computer Organization & Assembly Language © by DR. M. Amer.
Introduction to Research 2011 Introduction to Research 2011 Ashok Srinivasan Florida State University Images from ORNL, IBM, NVIDIA.
Parallelizing Spacetime Discontinuous Galerkin Methods Jonathan Booth University of Illinois at Urbana/Champaign In conjunction with: L. Kale, R. Haber,
Outline Why this subject? What is High Performance Computing?
Chapter VI What should I know about the sizes and speeds of computers?
DR. SIMING LIU SPRING 2016 COMPUTER SCIENCE AND ENGINEERING UNIVERSITY OF NEVADA, RENO Session 3 Computer Evolution.
Tackling I/O Issues 1 David Race 16 March 2010.
Hardware Trends CSE451 Andrew Whitaker. Motivation Hardware moves quickly OS code tends to stick around for a while “System building” extends way beyond.
Why Parallel/Distributed Computing Sushil K. Prasad
SPRING 2012 Assembly Language. Definition 2 A microprocessor is a silicon chip which forms the core of a microcomputer the concept of what goes into a.
Introduction to Computers - Hardware
High Performance Computing (HPC)
CMSC 611: Advanced Computer Architecture
Clouds , Grids and Clusters
for the Offline and Computing groups
CLUSTER COMPUTING Presented By, Navaneeth.C.Mouly 1AY05IS037
Nat 4/5 Computing Science Types of Computer and Performance
Hardware September 19, 2017.
Architecture & Organization 1
Vocabulary Big Data - “Big data is a broad term for datasets so large or complex that traditional data processing applications are inadequate.”
Nat 4/5 Computing Science Types of Computer and Performance
What is Parallel and Distributed computing?
Architecture & Organization 1
Hook was a famous year for what…..?
Parallel Processing Sharing the load.
Course Description: Parallel Computer Architecture
CSE8380 Parallel and Distributed Processing Presentation
CS 704 Advanced Computer Architecture
Vrije Universiteit Amsterdam
Presentation transcript:

UCSD SAN DIEGO SUPERCOMPUTER CENTER 1 Who needs a supercomputer? Professor Snavely, University of California Professor Allan Snavely University of California, San Diego and San Diego Supercomputer Center

Performance Modeling and Characterization Lab PMaC Arent computers fast enough already? This talk argues computers are not fast enough already Nor do supercomputers just naturally get faster as a result of Moores Law. We explore implications of: Moores Law Amdahls Law Einsteins Law Supercomputers are of strategic importance, enabling a Third Way of doing science-by-simulation Example: Terashake Earthquake simulation Viable National Cyberinfrastructure requires centralized supercomputers Supercomputing in Japan, Europe, India, China Why + Moores Law does not solve all our problems

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC The basic components of a computer Your laptop has these:

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Supercomputers (citius, altius, fortius) Supercomputers are just faster, higher, stronger, than your laptop, more and faster processors etc. capable of solving large scientific calculations

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC An army of ants approach In Supercomputers such as Blue Gene, DataStar, thousands of CPUs cooperate to solve scientific calculations

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Computers live a billion seconds to our every one! Definitions: Latency is distance measured in time Bandwidth is volume per unit of time Thus, in their own sense of time, the latencies and bandwdiths across the machine room span 11 orders of magnitude! (from Nanoseconds to Minutes.) To a supercomputer, getting data from disk is like sending a rocket- ship to Saturn!

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Moores Law Gordon Moore (co-founder of Intel) predicted in 1965 that the transistor density of semiconductor chips would double roughly every 18 months. Moores law has had a decidedly mixed impact, creating new opportunities to tap into exponentially increasing computing power while raising fundamental challenges as to how to harness it effectively. Things Moore never said: computers double in speed every 18 months cost of computing is halved every 18 months cpu utilization is halved every 18 months

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Moores Law Moores Law: the number of transistors per processor chip by doubles every 18 months.

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Snavelys Top500 Laptop? Among other startling implications of Moores Law is the fact that the peak performance of the typical laptop would have placed it as one of the 500 fastest computers in the world as recently as Shouldnt I just go find another job now? No, because Moores Law has several more subtle implications and these have raised a series of challenges to utilizing the apparently ever-increasing availability of compute power; these implications must be understood to see where we are today in High Performance superComputing (HPC).

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC The Vonn Neumann bottleneck Scientific calculations involve operations upon large amounts of data, and it is in moving data around within the computer that the trouble begins. As a very simple pedagogical example consider the expression A + B = C The computer has to load A and B, + them together, and store C + is fast by Moores Law, load and store is slow by Einsteins Law

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Supercomputer Red Shift While the absolute speed of all computer subcomponents have been changing rapidly, they have not all been changing at the same rate. While CPUs get faster they spend more time sitting around waiting for data

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Amdahls Law The law of diminishing returns When a task has multiple parts, after you speed up one part a lot, the other parts come to dominate the total time An example from cycling: On a hilly closed-loop course you cannot ever average more than 2x your uphill speed even if you go downhill at the speed of light! For supercomputers this means even though processors get faster the overall time to solution is limited by memory and interconnect speeds (moving the data around)

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Red Shift and the Red Queen It takes all the running you can do, to keep in the same place. If you want to get somewhere else, you must run at least twice as fast as that! Corollary: Allans laptop is not a balanced system! System utilization is cut in half every 18 months? Fundamental R&D in latency hiding, high bandwidth network, Computer Architecture

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC 3 ways of science Experiment Theory Simulation

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Major Earthquakes on the San Andreas Fault, 1680-present 1906 M M 7.7 The SCEC TeraShake simulation is a result of immense effort from the Geoscience community for over 10 years Focus is on understanding big earthquakes and how they will impact sediment- filled basins. Simulation combines massive amounts of data, high-resolution models, large-scale supercomputer runs TeraShake results provide new information enabling better Estimation of seismic risk Emergency preparation, response and planning Design of next generation of earthquake-resistant structures Such simulations provide potentially immense benefits in saving both many lives and billions in economic losses ? How Dangerous is the Southern San Andreas Fault?

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC TeraShake Animation

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Compute (more FLOPS) Data (more BYTES) Home, Lab, Campus, Desktop Traditional HPC environment Data-oriented Science and Engineering Environment SDSC and Data Intensive Computing Brain mapping TeraShake

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC The Japanese Earth Simulator Took U.S. HPC Community by surprise in 2002 – Computenik For 2 years had more flops capacity than top 5 U.S. systems Approach based on specialized HPC design Still has more data moving capacity Sparked space race in HPC, Blue Gene surpassed for flops 2005

San Diego Supercomputer Center Performance Modeling and Characterization Lab PMaC Summary Red Shift means the promise implied by Moores Law is largely unrealized for scientific simulation that by necessity operate on large data Consider The Butterfly Effect Supercomputer Architecture is a hot field Challenges from Japan, Europe, India, China Large centralized, specialized compute engines are a vital national strategic resources Grids, utility programing, etc. do not meet all the needs of largescale scientific simulation for reason that should now be Consider a galactic scale