30 Years of Navy Modeling and Supercomputers: an Anecdotal History Tom Rosmond Marine Meteorology Division Naval Research Laboratory Monterey, California.

Slides:



Advertisements
Similar presentations
Gordon Bell Bay Area Research Center Microsoft Corporation
Advertisements

SPEC High Performance Group (An Introduction)
L3S Research Center University of Hanover Germany
Clusters, Grids and their applications in Physics David Barnes (Astro) Lyle Winton (EPP)
SPEC ENV2002: Environmental Simulation Benchmarks Wesley Jones
Höchstleistungsrechenzentrum Stuttgart Matthias M üller SPEC Benchmarks at HLRS Matthias Mueller High Performance Computing Center Stuttgart
Eine Zeitreise in die Welt der Computer.1 Pipelined Vector Processing and Scientific Computation John G. Zabolitzky.
Weather Research & Forecasting: A General Overview
Interactive lesson about operating system
Opening Workshop DAS-2 (Distributed ASCI Supercomputer 2) Project vrije Universiteit.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Introduction to Computer Administration Introduction.
Operating Systems An Introduction. 2 What Does An Operating System Do? Manages the hardware and software resources of the system. In a desktop computer,
Super Computers By Phuong Vo.
Chapter 3 General-Purpose Processors: Software
Introductions to Parallel Programming Using OpenMP
PRIYADHARSHINI S SUPERCOMPUTERS. OVERVIEW  The term is commonly applied to the fastest high-performance systems in existence at the time of their construction.
Supercomputing Challenges at the National Center for Atmospheric Research Dr. Richard Loft Computational Science Section Scientific Computing Division.
Definitions of Supercomputer A time dependent term which refers to the class of most powerful computer systems world-wide at the time of reference An.
Beowulf Supercomputer System Lee, Jung won CS843.
Zhao Lixing.  A supercomputer is a computer that is at the frontline of current processing capacity, particularly speed of calculation.  Supercomputers.
Types of Parallel Computers
History of Distributed Systems Joseph Cordina
Introduction CS 524 – High-Performance Computing.
1 WRF Development Test Center A NOAA Perspective WRF ExOB Meeting U.S. Naval Observatory, Washington, D.C. 28 April 2006 Fred Toepfer NOAA Environmental.
Supercomputers Daniel Shin CS 147, Section 1 April 29, 2010.
NPACI: National Partnership for Advanced Computational Infrastructure Supercomputing ‘98 Mannheim CRAY T90 vs. Tera MTA: The Old Champ Faces a New Challenger.
Introduction What is Parallel Algorithms? Why Parallel Algorithms? Evolution and Convergence of Parallel Algorithms Fundamental Design Issues.
1 New Architectures Need New Languages A triumph of optimism over experience! Ian Watson 3 rd July 2009.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Programming for High Performance Computers John M. Levesque Director Cray’s Supercomputing Center Of Excellence.
Operational computing environment at EARS Jure Jerman Meteorological Office Environmental Agency of Slovenia (EARS)
OpenMP in a Heterogeneous World Ayodunni Aribuki Advisor: Dr. Barbara Chapman HPCTools Group University of Houston.
II.B.2 COMPUTATIONAL ISSUES and CONSIDERATIONS for DA TOM ROSMOND Forks, Washington (SAIC) JCSDA Summer Colloquium on Data Assimilation Santa Fe, New Mexico.
Lappeenranta University of Technology / JP CT30A7001 Concurrent and Parallel Computing Introduction to concurrent and parallel computing.
Reconfigurable Devices Presentation for Advanced Digital Electronics (ECNG3011) by Calixte George.
Distributed Shared Memory: A Survey of Issues and Algorithms B,. Nitzberg and V. Lo University of Oregon.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Classification of Computers
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
The WRF Model The Weather Research and Forecasting (WRF) Model is a mesoscale numerical weather prediction system designed for both atmospheric research.
AN EXTENDED OPENMP TARGETING ON THE HYBRID ARCHITECTURE OF SMP-CLUSTER Author : Y. Zhao 、 C. Hu 、 S. Wang 、 S. Zhang Source : Proceedings of the 2nd IASTED.
Experience with COSMO MPI/OpenMP hybrid parallelization Matthew Cordery, William Sawyer Swiss National Supercomputing Centre Ulrich Schättler Deutscher.
Angèle Simard Canadian Meteorological Center Meteorological Service of Canada MSC Computing Update.
Frank Casilio Computer Engineering May 15, 1997 Multithreaded Processors.
 Copyright, HiCLAS1 George Delic, Ph.D. HiPERiSM Consulting, LLC And Arney Srackangast, AS1MET Services
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Gouge – Navy slang for the bare essential knowledge to get by.
March, 2003 SOS 7 Jim Harrell Unlimited Scale Inc.
Thirty Years of Operational Ocean Wave Forecasting at Fleet Numerical Meteorology and Oceanography Center Paul Wittmann and Mike Clancy Fleet Numerical.
Computing Environment The computing environment rapidly evolving ‑ you need to know not only the methods, but also How and when to apply them, Which computers.
Parallel Computing.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
Outline Why this subject? What is High Performance Computing?
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
Background Computer System Architectures Computer System Software.
Constructing a system with multiple computers or processors 1 ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson. Jan 13, 2016.
Keith Kelley Presentation 2 Weather Forecasting with Parallel Computers.
Group Members Hamza Zahid (131391) Fahad Nadeem khan Abdual Hannan AIR UNIVERSITY MULTAN CAMPUS.
Conclusions on CS3014 David Gregg Department of Computer Science
Auburn University COMP8330/7330/7336 Advanced Parallel and Distributed Computing Parallel Hardware Dr. Xiao Qin Auburn.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Green cloud computing 2 Cs 595 Lecture 15.
Super Computing By RIsaj t r S3 ece, roll 50.
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
Types of Parallel Computers
Presentation transcript:

30 Years of Navy Modeling and Supercomputers: an Anecdotal History Tom Rosmond Marine Meteorology Division Naval Research Laboratory Monterey, California USA

Outline Before my Time: My Early Years: The Golden Years of NWP: late 1970’s-early 1980’s Up to the Present Conclusions and Thoughts for the Future

Before my Time: : Capt Wolff sets up shop on NPS Campus, Monterey CDC 1604: Seymour Cray’s first design with CDC (first supercomputer?) 1961: Establishment of Fleet Numerical Weather Facility (FNWF) 1964: Routine dissemination of numerical products Successive correction analyses of SLP, upper air heights NH Barotropic and thickness advection models THE BEGINNING

Before my Time: : FNWC procurement of CDC 6500 to replace 1604 Dual-processors, comparable to 386 microprocessor in FP performance Supercomputer of time was Seymour Cray’s CDC 6600 (NMC in 1966) 1969: Second 6500 acquired 1970: Northern Hemisphere Primitive Equation (NHPE) model operational on 4 processors. World’s first multi-processor production code CDC 6500’s and NHPE

Before my Time: Developed by P. Kessel and F. Winninghoff Processors shared model data through extended core storage (ECS) OS modifications to allow processor synchronization Similarities to both shared memory (OpenMP) and distributed memory (MPI) programming models Parallel efficiency = ~ 75% An impressive achievement that is certainly underappreciated An example of the struggle it was to fit models onto these early systems 4-processor 6500 NHPE

My Early Years: My organization: EPRF NEPRF NOARL NRL-Monterey 1974: I joined the Environmental Prediction Research Facility (EPRF) 1 Little involvement in FNWC 2 model development Access of FNWC computer systems (6500’s, 1604) 1976: NEPRF Numerical Modeling Department formed at request of FNWC CO Capt. R. Hughes He realized that FNWC was unable to maintain R & D continuity needed for NWP system development. Need for a global NWP capability was major motivation, and he was committed to getting computer system to support it. 1976: FNWC acquired CDC CYBER 175 (similar to CDC 7600) 2 Fleet Numerical: FNWF FNWC FNOC FNMOC

My Early Years: NEPRF Global NWP System Development 1976: Navy Operational Global Atmospheric Prediction System (NOGAPS) UCLA General Circulation Model Barnes Successive Corrections Global Analysis Variational Initialization with Balance Equation Constraint 1977: We also had spectral model dynamical core UCLA GCM physics Sat on shelf for several years in favor of UCLA GCM based system 1980: Prototype NOGAPS running on CYBER 175

My Early Years: : Benchmarking for FNWC procurement NCAR Cray-1, first system outside Los Alamos? UCLA GCM x 3.0 0, 6 levels First introduction to vector programming Crude compiler + non-vectorizable code = poor performance Overly conservative target performance, CYBER 203 could compete Allowed subsequent CDC success with CYBER : CYBER 203 delivered to FNOC Heroic work by CDC to get UCLA model to run fast 1982: CYBER 203 replaced with CYBER 205

The Golden Years of NWP 1 st half: late 70’s – mid 80’s Supercomputers were cheapest computers you could buy (price/performance) Cray: Cray-1, XMP, etc CDC: CYBER 203/205, ETA-10 IBM: VP Development of global NWP forecast systems Dominance of spectral models Increased realization of importance of data assimilation Establishment of ECMWF “Raised the stakes” in operational NWP Accelerated progress Provided gold standard

The Golden Years of NWP Navy operational models NOGAPS UCLA-GCM x3.00 x L9 : 1-pipe/8Mbyte C205 : ~25min/fcstday Spectral – T47L18 : 2-pipe/32Mbyte C205 : ~6min/fcstday Spectral – T79L18 : 2-pipe/32Mbyte C205 : ~25min/fcstday (32bit) Spectral – T79L24 : 4-pipe/64Mbyte C205 : ~12min/fcstday (32bit) NORAPS (developed by Rich Hodur, NEPRF) Globally relocatable regional model 6-8 areas run operationally Varying resolution, domain and grid sizes e.g. ran over South Atlantic during Falklands war (1980)

Some special comments about the CYBER 205 CDC aggressively pursued meteorology market, both operations and research FNOC, NMC, UKMO, GSFC, GFDL, etc CRAY systems were better general purpose, but 205 excelled on our applications Initially 205 was difficult to use, compiler/user software was rudimentary, but Language extension showed how machine worked Rich array of exotic vector hardware instructions System software matured 32 bit/64 bit floating point support Easy mixing of Fortran and explicit vector instructions, we could get “close to the hardware” Spectacular percentage of peak performance possible

The Golden Years of NWP 2 nd half: mid 80’s – early 90’s Introduction of multi-processor systems Divide and conquer X/MP, Y/MP, ETA-10 Multi-tasking/vectorization programming model (parallel/vector) Spelled the end of single processor supercomputers, e.g. CYBER 205 Important changes in technology Price/performance advantage of supercomputers ending Introduction of desktop workstations Many people didn’t need supercomputers, just cheap computing Shrinking supercomputer market

Up to the Present: early 1990’s 1990: Introduction of CRAY C90 The “best” supercomputer ever? Run by a higher percentage of NWP operational/research centers than any system before or since. Parallel/vector programming model very user friendly Easy to get high percentage of peak performance 1991: 8 processor C90 at FNMOC NOGAPS: T159L24 – 6 processors : 10 min/fcstday 1996: 8 and16 processor C90’s at FNMOC NOGAPS: T159L24 – 12 processors : 6 min/fcstday 1997: COAMPS TM replaced NORAPS as Navy’s regional forecast system

Up to the Present Mid to Late 1990’s Beginnings of dramatic changes in supercomputer industry Scalable commodity based architectures appearing, e.g. CRAY T3E Powerful workstations replacing supercomputers for many applications End of domination by American supercomputer vendors CDC/ETA were long gone CRAY sold to SGI MTTB (mean time to bankruptcy) very short for new companies Rise of Japanese vendors to dominate NWP marketplace, at least outside U.S. Fujitsu NEC

Up to the Present 2000 to today Proliferation of commodity based, scalable architectures for NWP applications Many T3E’s still in use SGI: Origin 2000, Origin 3000 IBM: SP3, SP4 Linux clusters are also viable alternatives, especially when price/performance is overriding issue But, rumors of the demise of vector architectures are exaggerated 2000: NEC only vendor with traditional vector architecture, SX-6, but 2002: Resurrected CRAY, Inc introduced X-1 Price/performance vs ultimate performance is central question concerning long-term prospects of these systems

Up to the Present 2000 to today 2001: FNMOC replaced C90’s with O3000’s 1152 total processors NOGAPS: T239L30 – 120 processors: 6 min/fcstday COAMPS TM : 8-11 areas – processors: min/fcstday 2003: Operational NAVDAS/NOGAPS (3DVAR): 60 processors 2004: Direct radiance retrievals with NAVDAS 2004: Currently under development at NRL, Monterey T479L54 semi-Lagrangian NOGAPS NAVDAS-AR: 4DVAR extension of NAVDAS COAMPS TM /WRF: more areas, higher resolution Clearly computational requirements are never satisfied!!!

Conclusions and thoughts for the future Vector architectures will continue to be viable candidates for NWP applications Sharing some features with commodity systems, e.g. caches When ultimate performance overrides price/performance Will be part of heterogeneous computing environments Scientific computing ( and therefore NWP) is now niche market Be thankful that video game applications share many of our requirements Consumer based industry has driven some hardware costs to astonishingly low levels, e.g. hard drives We must work with vendors to ensure that our requirements are not forgotten Fortran compilers High-performance interconnects