DoD HPC Modernization Program & Move Toward Emerging Architectures

Slides:



Advertisements
Similar presentations
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
Advertisements

The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
1 Computational models of the physical world Cortical bone Trabecular bone.
©2009 HP Confidential template rev Ed Turkel Manager, WorldWide HPC Marketing 4/7/2011 BUILDING THE GREENEST PRODUCTION SUPERCOMPUTER IN THE.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
Appro Xtreme-X Supercomputers A P P R O I N T E R N A T I O N A L I N C.
Contact: Hirofumi Amano at Kyushu 40 Years of HPC Services In this memorable year, the.
Parallel Research at Illinois Parallel Everywhere
Why GPU Computing. GPU CPU Add GPUs: Accelerate Science Applications © NVIDIA 2013.
Cyberinfrastructure for Scalable and High Performance Geospatial Computation Xuan Shi Graduate assistants supported by the CyberGIS grant Fei Ye (2011)
HPC at University of Moratuwa & Sri Lanka Dilum Bandara, PhD Senior Lecturer Dept. of Computer Science & Engineering, University of.
ASKAP Central Processor: Design and Implementation Calibration and Imaging Workshop 2014 ASTRONOMY AND SPACE SCIENCE Ben Humphreys | ASKAP Software and.
Claude TADONKI Mines ParisTech – LAL / CNRS / INP 2 P 3 University of Oujda (Morocco) – October 7, 2011 High Performance Computing Challenges and Trends.
Heterogeneous Computing: New Directions for Efficient and Scalable High-Performance Computing Dr. Jason D. Bakos.
1 Maui High Performance Computing Center Open System Support An AFRL, MHPCC and UH Collaboration December 18, 2007 Mike McCraney MHPCC Operations Director.
Dell IT Innovation Labs in the Cloud “The power to do more!” Andrew Underwood – Manager, HPC & Research Computing APJ Solutions Engineering Team.
Lecture 1: Introduction to High Performance Computing.
1 AppliedMicro X-Gene ® ARM Processors Optimized Scale-Out Solutions for Supercomputing.
Aim High…Fly, Fight, Win NWP Transition from AIX to Linux Lessons Learned Dan Sedlacek AFWA Chief Engineer AFWA A5/8 14 MAR 2011.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
GPU Programming with CUDA – Accelerated Architectures Mike Griffiths
High Performance Computing G Burton – ICG – Oct12 – v1.1 1.
© Fujitsu Laboratories of Europe 2009 HPC and Chaste: Towards Real-Time Simulation 24 March
SDSC RP Update TeraGrid Roundtable Reviewing Dash Unique characteristics: –A pre-production/evaluation “data-intensive” supercomputer based.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
HPC Technology Track: Foundations of Computational Science Lecture 1 Dr. Greg Wettstein, Ph.D. Research Support Group Leader Division of Information Technology.
The WRF Model The Weather Research and Forecasting (WRF) Model is a mesoscale numerical weather prediction system designed for both atmospheric research.
Export Controls—What’s next? Joseph Young Bureau of Industry and Security Export Controls – What’s Next? Joseph Young Bureau of Industry and Security.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Are Supercomputers returning to Investment Banking?
Early Experiences with Energy-Aware (EAS) Scheduling
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
HPC Business update HP Confidential – CDA Required
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Contact: Hirofumi Amano at Kyushu Mission 40 Years of HPC Services Though the R. I. I.
HPCMP Benchmarking Update Cray Henry April 2008 Department of Defense High Performance Computing Modernization Program.
1 Latest Generations of Multi Core Processors
International Exascale Software Co-Design Workshop: Initial Findings & Future Plans William Tang Princeton University/Princeton Plasma Physics Lab
Introduction to Research 2011 Introduction to Research 2011 Ashok Srinivasan Florida State University Images from ORNL, IBM, NVIDIA.
© 2009 IBM Corporation Motivation for HPC Innovation in the Coming Decade Dave Turek VP Deep Computing, IBM.
RECAPITALIZING THE NATION’S WEATHER PREDICTION CAPABILITY National Unified Operational Prediction Capability (NUOPC)
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
Program Update Presented by Larry Davis, Deputy Director September 2009 Department of Defense High Performance Computing Modernization Program.
IDC HPC User Forum April 14 th, 2008 A P P R O I N T E R N A T I O N A L I N C Steve Lyness Vice President, HPC Solutions Engineering
NICS RP Update TeraGrid Round Table March 10, 2011 Ryan Braby NICS HPC Operations Group Lead.
High Performance Computing
ISI Research Opportunities at Maui High Performance Computing Center ISI Research Opportunities at Maui High Performance Computing Center Developed for.
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
Presented by NCCS Hardware Jim Rogers Director of Operations National Center for Computational Sciences.
Seaborg Decommission James M. Craw Computational Systems Group Lead NERSC User Group Meeting September 17, 2007.
Moore’s Law Electronics 19 April Moore’s Original Data Gordon Moore Electronics 19 April 1965.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & LSU Condor in Louisiana Tevfik Kosar Center for Computation & Technology Louisiana.
Scheduling a 100,000 Core Supercomputer for Maximum Utilization and Capability September 2010 Phil Andrews Patricia Kovatch Victor Hazlewood Troy Baer.
Petascale Computing Resource Allocations PRAC – NSF Ed Walker, NSF CISE/ACI March 3,
The Evolution of the Italian HPC Infrastructure Carlo Cavazzoni CINECA – Supercomputing Application & Innovation 31 Marzo 2015.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Hopper The next step in High Performance Computing at Auburn University February 16, 2016.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Earth System Modelling: an HPC perspective Mike Ashworth & Rupert Ford Scientific Computing Department and STFC Hartree Centre STFC Daresbury Laboratory.
Sobolev(+Node 6, 7) Showcase +K20m GPU Accelerator.
Modern supercomputers, Georgian supercomputer project and usage areas
Appro Xtreme-X Supercomputers
HPCMP New Users’ Guide “Who Are We?”
Scientific Computing At Jefferson Lab
Title of Presentation Client
Footer.
K computer RIKEN Advanced Institute for Computational Science
CUBAN ICT NETWORK UNIVERSITY COOPERATION (VLIRED
K computer RIKEN Advanced Institute for Computational Science
Presentation transcript:

DoD HPC Modernization Program & Move Toward Emerging Architectures Tom Dunn Naval Meteorology & Oceanography Command 20 November 2014

HPC RECENT TRENDS Per Top500 List RECENT 2014 DOD ACQUISITIONS EXPECTED PROCESSOR COMPETITION ONWARD TOWARD EXASCALE

Navy DoD SUPERCOMPUTING RESOURCE CENTER Peak Computational Performance (Teraflops) Estimates Follow Moore’s Law (~2x every 2 yrs) 1997 – .3 TFs 2012 (Dec) – 954 TFs 2001 – 8.4 TFs 2014 (Jul) – 2,556 TFs 2004 - 32 TFs 2015 (Jul) – 5,760 TFs est. 2006 – 58 TFs 2017 (Jul) –10,000 TFs est. 2008 – 226 TFs 3

Navy DSRC Capabilities One of the most capable HPC centers in the DoD and the nation Chartered as a DoD Supercomputing Center in 1994 Computational performance approximately doubles every two years; Currently 2,556 Teraflops Systems reside on the Defense Research and Engineering Network (DREN) with 10 Gb connectivity – 19 Dec 2013 15% of Navy DSRC’s computational and storage capacity reserved for CNMOC activities operational use R&D and CNMOC Ops are placed in separate system partitions and queues 4

Top500® Systems by Architecture, June 2006–June 2014

Number of CPUs in the Top500® Systems by Architecture Type, June 2006–June 2014

Number of Systems in the Top500® Utilizing Co-Processors or Accelerators, June 2009–June 2014 7

Number of Systems in the Top500® by Co-Processors or Accelerators Type, June 2009–June 2014 8

Number of Cores in the Top500® by Co-Processors or Accelerators Type, June 2011–June 2014 9

Number of Cores in the June 2014 Top500® by CPU Manufacturer 10

TOP 500 SUPERCOMPUTER LIST (JUNE 2014) BY OEM Supplier TOP 500 CRAY INC 51 DELL 8 HEWLETT PACKARD 182 IBM 176 SGI 19 TOTAL 436 Other Suppliers 64

High Performance Computing Modernization Program 2014 HPC Awards Feb. 2014 Air Force Research Lab (AFRL) DSRC, Dayton, OH Cray XC-30 System (Lightning) - 1281 teraFLOPS - 56,880 Compute Cores (2.7 GHz Intel Ivy Bridge) - 32 NVIDIA Tesla K40 GPGPUs Navy DSRC, Stennis Space Center, MS Cray XC-30 (Shepard) - 813 teraFLOPS - 28,392 Compute Cores (2.7 GHz Intel Ivy Bridge) - 124 Hybrid nodes, each consisting of 10 Ivy Bridge cores and a 60 core Intel Xeon 5120D Phi Cray XC-30 (Armstrong) - 786 teraFLOPS - 29,160 Compute cores (2.7 GHz Intel Ivy Bridge) 12

High Performance Computing Modernization Program 2014 HPC Awards September 2014 Army Research Lab (ARL) DSRC, Aberdeen, MD Cray XC-40 System - 3.77 petaFLOPS - 101,312 compute cores (2.3 GHz Intel Xeon Haswell) - 32 NVIDIA Tesla K40 GPGPUs - 411 TB memory - 4.6 PB storage Army Engineer Research Development Center (ERDC) DSRC, Vicksburg, MS SGI ICE X System - 4.66 petaFLOPS - 125,440 compute cores (2.3 GHz Intel Xeon Haswell) - 32 NVIDIA Tesla K40 GPGPUs - 440 TB memory - 12.4 PB storage 13

High Performance Computing Modernization Program 2014/2015 HPC Awards Air Force Research Lab (AFRL) DSRC, Dayton, OH FY15 Funded OEM and Contract Award - TBD - 100,000+ compute cores - 3.5 – 5.0 petaFLOPS Navy DSRC, Stennis Space Center, MS 14

Weather & Climate Operational Supercomputing System (WCOSS) ECMWF (Top 500 List Jun 2014) 2 Cray XC30 Systems each with 81,160 compute cores (2.7 GHz Intel Ivy Bridge) 1,796 teraFLOPS NOAA NWS/NCEP Weather & Climate Operational Supercomputing System (WCOSS) Phase I 2 IBM iDataplex systems each with 10,048 compute cores (2.6 GHz Intel Sandy Bridge) 213 teraFLOPS Phase II (Jan 2015) Addition 2 IBM NeXtScale systems each with 24,192 compute cores (2.7GHz Intel Ivy Bridge) 585 teraFLOPS 15

UK Meterological Office IBM Power 7 System 18,432 compute cores (3 UK Meterological Office IBM Power 7 System 18,432 compute cores (3.836 GHz) 565 teraFLOPS 15,360 compute cores (3.876 GHz) 471 teraFLOPS ---------------------------------------------------------------------------------------------------- 27 Oct 2014 Announcement 128M Contract 2 Cray XC-40 systems (Intel Xeon Haswell initially) >13 times faster than current system total of 480,000 compute cores Phase 1a replace Power 7s by Sep 2015 Phase 1b extend both systems to power limit by Mar 2016 Phase 1c add one new system by Mar 2017 16

Expected Near Term HPC Processor Options 2016 Intel and ARM - Cray has ARM in-house for testing 2017 - Intel, ARM, & IBM Power 9 (with closely coupled NVIDIA GPUs) 17

DoD Applications & Exascale Computing General external impression In the 2024 timeframe, DoD will have no requirement for a balanced exascale supercomputer (untrue) DoD should not be a significant participant in exascale planning for the U.S. (untrue) Reality DoD has compelling coupled multi-physics problems which will require more tightly-integrated resources than technologically possible in the 2024 timeframe DoD has many other use cases which will benefit from the power efficiencies and novel technologies generated by the advent of exascale computing 18

HPCMP & 2024 DoD Killer Applications   HPCMP Categorizes Users Base into 11 Computational Technology Areas (CTAs) Climate Weather Ocean (CWO) is one of 11 CTAs Dr. Burnett (CNMOC TD) is the DoD HPCMP CWO CTA leader Each CTA leader tasked in FY14 to project Killer Apps in their CTA Dr. Burnett’s CWO CTA analysis lead by Lockheed Martin Primary focus is on HYCOM but includes NAVGEM, and ESPC Expect follow-on FY15 funding Develop appropriate Kiviat diagrams (example to follow) NRL Stennis part of an ONR sponsored NOPP project starting FY14 to look at attached processors (i.e. GPGPUs and accelerators) for HYCOM+CICE+WW3 19

Relevant Technology Issues Classical computing advances may stall in the next 10 years 22nm (feature size for latest processors) 14nm (anticipated feature size in 2015) 5-7nm (forecast limit for classical methods) Recent 3D approaches currently used and dense 3D approaches contemplated, but have limitations Mean-time-between-failures (MTBF) will decrease dramatically Petascale (hours to days) Exascale (minutes) Data management exacale hurdles Power management exascale hurdles 20

Relevant Software Issues Gap between intuitive coding (i.e. readily relatable to domain science) and high performance coding will increase Underpinnings of architectures will change more rapidly than codes can be refactored Parallelism of underlying mathematics will become asymptotic (at some point) despite the need to scale to millions [if not billions] of processing cores Current parallel code is based (in general) on synchronous communications; however, asynchronous methods may be necessary to overcome technology issues 21

Path Forward (Deliverables) [cont.] Kiviat diagram conveying system architecture requirements for each impactful advent 22

March Toward Exascale Computing Dept of Energy target for exascale in 2024 Japan target for exascale in 2020 (with $1B gov assistance) China target for exascale now in 2020 (originally in 2018) HPCMP’s systems expected in 7 or 8 years – 100 petaflops 23