Brent Gorda LBNL – SOS7 3/5/03 1 Planned Machines: BluePlanet SOS7 March 5, 2003 Brent Gorda Future Technologies Group Lawrence Berkeley.

Slides:



Advertisements
Similar presentations
Discussion of Infrastructure Clouds A NERSC Magellan Perspective Lavanya Ramakrishnan Lawrence Berkeley National Lab.
Advertisements

O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
Slides Prepared from the CI-Tutor Courses at NCSA By S. Masoud Sadjadi School of Computing and Information Sciences Florida.
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Blue Gene/P System Overview - Hardware.
U.S. Department of Energy’s Office of Science Basic Energy Sciences Advisory Committee Dr. Daniel A. Hitchcock October 21, 2003
♦ Commodity processor with commodity inter- processor connection Clusters Pentium, Itanium, Opteron, Alpha GigE, Infiniband, Myrinet, Quadrics, SCI NEC.
Today’s topics Single processors and the Memory Hierarchy
Beowulf Supercomputer System Lee, Jung won CS843.
Planned Machines: ASCI Purple, ALC and M&IC MCR Presented to SOS7 Mark Seager ICCD ADH for Advanced Technology Lawrence Livermore.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
1 NERSC User Group Business Meeting, June 3, 2002 High Performance Computing Research Juan Meza Department Head High Performance Computing Research.
Energy Efficient Prefetching – from models to Implementation 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software Engineering.
Building a Cluster Support Service Implementation of the SCS Program UC Computing Services Conference Gary Jung SCS Project Manager
Hitachi SR8000 Supercomputer LAPPEENRANTA UNIVERSITY OF TECHNOLOGY Department of Information Technology Introduction to Parallel Computing Group.
Arquitectura de Sistemas Paralelos e Distribuídos Paulo Marques Dep. Eng. Informática – Universidade de Coimbra Ago/ Machine.
1 Computer Science, University of Warwick Architecture Classifications A taxonomy of parallel architectures: in 1972, Flynn categorised HPC architectures.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Simulating Quarks and Gluons with Quantum Chromodynamics February 10, CS635 Parallel Computer Architecture. Mahantesh Halappanavar.
LLNL-PRES-XXXXXX This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program CASC, May 3, ADVANCED SCIENTIFIC COMPUTING RESEARCH An.
Network Aware Resource Allocation in Distributed Clouds.
Integrating Fine-Grained Application Adaptation with Global Adaptation for Saving Energy Vibhore Vardhan, Daniel G. Sachs, Wanghong Yuan, Albert F. Harris,
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Seaborg Cerise Wuthrich CMPS Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture.
R. Ryne, NUG mtg: Page 1 High Energy Physics Greenbook Presentation Robert D. Ryne Lawrence Berkeley National Laboratory NERSC User Group Meeting.
CSIE30300 Computer Architecture Unit 15: Multiprocessors Hsin-Chou Chi [Adapted from material by and
Crystal Ball Panel ORNL Heterogeneous Distributed Computing Research Al Geist ORNL March 6, 2003 SOS 7.
ESIP Federation Air Quality Cluster Partner Agencies.
A High-Performance Scalable Graphics Architecture Daniel R. McLachlan Director, Advanced Graphics Engineering SGI.
March 9, 2015 San Jose Compute Engineering Workshop.
Contact: Hirofumi Amano at Kyushu Mission 40 Years of HPC Services Though the R. I. I.
1 CMPE 511 HIGH PERFORMANCE COMPUTING CLUSTERS Dilek Demirel İşçi.
1 HPC as Service and Scientific Applications Bruno Schulze.
CCS Overview Rene Salmon Center for Computational Science.
Co-Design 2013 Summary Exascale needs new architectures due to slowing of Dennard scaling (since 2004), multi/many core limits New programming models,
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
Cray Environmental Industry Solutions Per Nyberg Earth Sciences Business Manager Annecy CAS2K3 Sept 2003.
1 DOE Office of Science October 2003 SciDAC Scientific Discovery through Advanced Computing Alan J. Laub.
Mellanox Connectivity Solutions for Scalable HPC Highest Performing, Most Efficient End-to-End Connectivity for Servers and Storage April 2010.
Enhancing Commodity Scalar Processors with Vector Components for Increased Scientific Productivity.
National Energy Research Scientific Computing Center (NERSC) HPC In a Production Environment Nicholas P. Cardo NERSC Center Division, LBNL November 19,
1 OFFICE OF ADVANCED SCIENTIFIC COMPUTING RESEARCH The NERSC Center --From A DOE Program Manager’s Perspective-- A Presentation to the NERSC Users Group.
Interconnection network network interface and a case study.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Exscale – when will it happen? William Kramer National Center for Supercomputing Applications.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
Presented by NCCS Hardware Jim Rogers Director of Operations National Center for Computational Sciences.
CDA-5155 Computer Architecture Principles Fall 2000 Multiprocessor Architectures.
Background Computer System Architectures Computer System Software.
1 Infrastructure Needs for Research and Development and Education Rational You ITRI-IEK-NEMS 2001/08/01 Source: IWGN (1999/09)
IBM Power system – HPC solution
VU-Advanced Computer Architecture Lecture 1-Introduction 1 Advanced Computer Architecture CS 704 Advanced Computer Architecture Lecture 1.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
BLUE GENE Sunitha M. Jenarius. What is Blue Gene A massively parallel supercomputer using tens of thousands of embedded PowerPC processors supporting.
Lecture 13 Parallel Processing. 2 What is Parallel Computing? Traditionally software has been written for serial computation. Parallel computing is the.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Centre of Excellence in Physics at Extreme Scales Richard Kenway.
Group Members Hamza Zahid (131391) Fahad Nadeem khan Abdual Hannan AIR UNIVERSITY MULTAN CAMPUS.
READ ME FIRST Use this template to create your Partner datasheet for Azure Stack Foundation. The intent is that this document can be saved to PDF and provided.
Clouds , Grids and Clusters
Web: Parallel Computing Rabie A. Ramadan , PhD Web:
The Earth Simulator System
BlueGene/L Supercomputer
CLUSTER COMPUTING.
Course Description: Parallel Computer Architecture
Defining the Grid Fabrizio Gagliardi EMEA Director Technical Computing
Presentation transcript:

Brent Gorda LBNL – SOS7 3/5/03 1 Planned Machines: BluePlanet SOS7 March 5, 2003 Brent Gorda Future Technologies Group Lawrence Berkeley National Lab

National Energy Research Scientific Computing Center ~2000 Users in ~400 projects Serves all disciplines of the DOE Office of Science NERSC Focus on large-scale computing

Brent Gorda LBNL – SOS7 3/5/03 3 The Divergence Problem The requirements of high performance computing for science and engineering, and the requirements of the commercial market are diverging. The commercial cluster of SMP approach is no longer sufficient to provide the highest level of performance –Lack of memory bandwidth - high memory latency –Lack of interconnect bandwidth - high interconnect latency –The commodity building block was the microprocessor but is now the entire server (SMP)! U.S. computer industry is driven by commercial applications The decision for NERSC-3 E can be seen as an indication of the divergence problem: Power 4 had a low SSP number – scaling problem

Brent Gorda LBNL – SOS7 3/5/03 4 Workshops: Sept 2002 – defining the Blue Planet architecture Nov – IBM gathered input for Power 6 White Paper: "Creating Science-Driven Computer Architecture: A New Path to Scientific Leadership,“ Cooperative Development – NERSC/ANL/IBM Workshop

Brent Gorda LBNL – SOS7 3/5/03 5 What is unique… … in the design process: Awareness and sensitivity to vendor’s mainstream products Get back “inside the box” … in the collaboration: Application teams drive the design process …in the structure and function of the system: Single core cpu design – not a previously planned product Additional Federation stage – 4x scale up in links Virtual Vector Architecture – ViVA

Brent Gorda LBNL – SOS7 3/5/03 6 Applications to Drive the Design of New Architectures Combustion Simulation and Adaptive Methods Computational Astrophysics Nanoscience Climate Modeling Accelerator Modeling Lattice Quantum Chromodynamics Quantum Monte Carlo Calculations of Nuclei High Energy / Elementary Particle Physics Biochemical and Biosystems Simulations Advanced Simulations of Plasma Microturbulence Computational Environmental Molecular Science Application’s needs vary widely, however most utilize MPI

Brent Gorda LBNL – SOS7 3/5/03 7 Interaction based on Scientific Apps AMRCoupled Climate AstrophysicsNanoscience MADCAPCactusFLAPWLSMS Sensitive to global bisection XXXX Sensitive to processor to memory latency XXX Sensitive to network latency XXXXX Sensitive to point to point communications XXX Sensitive to OS interference in frequent barriers XX Benefits from deep CPU pipelining XXXXXX Benefits from Large SMP nodes X

Brent Gorda LBNL – SOS7 3/5/03 8 Slide courtesy of Peter Ungaro, IBM

Brent Gorda LBNL – SOS7 3/5/03 9 What prior experience guided this choice? Power 4 memory bandwidth does not support 32 CPUs, and Power 4 Memory Latency is only 29% longer than Power 3.

Brent Gorda LBNL – SOS7 3/5/03 10 The majority of applications achieve low percentage of peak –Adding cpu’s w/o adding bandwidth makes no sense to HPC applications The majority of applications are data starved: memory and interconnect Many of the applications are regular and may be able to take advantage of ViVa What prior experience guided this choice?

Brent Gorda LBNL – SOS7 3/5/03 11 Other than your own machine, for your needs what are the best and worst machines? Difficult to push the envelope as a production facility Balance would appear to be the most important technical feature for a general purpose system Stability is the most important production feature  Both involve more than just hardware The best system is one that lives up to (sometimes slightly reduced) expectations