Toward a Unified HPC and Big Data Runtime

Slides:



Advertisements
Similar presentations
Issues of HPC software From the experience of TH-1A Lu Yutong NUDT.
Advertisements

Technology Drivers Traditional HPC application drivers – OS noise, resource monitoring and management, memory footprint – Complexity of resources to be.
Phillip Dickens, Department of Computer Science, University of Maine. In collaboration with Jeremy Logan, Postdoctoral Research Associate, ORNL. Improving.
1 A GPU Accelerated Storage System NetSysLab The University of British Columbia Abdullah Gharaibeh with: Samer Al-Kiswany Sathish Gopalakrishnan Matei.
Performance Metrics Inspired by P. Kent at BES Workshop Run larger: Length, Spatial extent, #Atoms, Weak scaling Run longer: Time steps, Optimizations,
DEPARTMENT OF COMPUTER LOUISIANA STATE UNIVERSITY Models without Borders Thomas Sterling Arnaud & Edwards Professor, Department of Computer Science.
Priority Research Direction (I/O Models, Abstractions and Software) Key challenges What will you do to address the challenges? – Develop newer I/O models.
Prof. Srinidhi Varadarajan Director Center for High-End Computing Systems.
1 Lawrence Livermore National Laboratory By Chunhua (Leo) Liao, Stephen Guzik, Dan Quinlan A node-level programming model framework for exascale computing*
Graph Analysis with High Performance Computing by Bruce Hendrickson and Jonathan W. Berry Sandria National Laboratories Published in the March/April 2008.
NGNS Program Managers Richard Carlson Thomas Ndousse ASCAC meeting 11/21/2014 Next Generation Networking for Science Program Update.
Basics of Operating Systems March 4, 2001 Adapted from Operating Systems Lecture Notes, Copyright 1997 Martin C. Rinard.
Bridge the gap between HPC and HTC Applications structured as DAGs Data dependencies will be files that are written to and read from a file system Loosely.
Leveraging Hierarchy Is this our Undiscovered Country? John T. Daly.
Fine Grain MPI Earl J. Dodd Humaira Kamal, Alan University of British Columbia 1.
Course Outline DayContents Day 1 Introduction Motivation, definitions, properties of embedded systems, outline of the current course How to specify embedded.
ET E.T. International, Inc. X-Stack: Programming Challenges, Runtime Systems, and Tools Brandywine Team May2013.
Scalable Data Clustering with GPUs Andrew D. Pangborn Thesis Defense Rochester Institute of Technology Computer Engineering Department Friday, May 14 th.
Tennessee Technological University1 The Scientific Importance of Big Data Xia Li Tennessee Technological University.
N. GSU Slide 1 Chapter 02 Cloud Computing Systems N. Xiong Georgia State University.
CCA Common Component Architecture Manoj Krishnan Pacific Northwest National Laboratory MCMD Programming and Implementation Issues.
Extreme scale parallel and distributed systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
Architectural Support for Fine-Grained Parallelism on Multi-core Architectures Sanjeev Kumar, Corporate Technology Group, Intel Corporation Christopher.
Introduction, background, jargon Jakub Yaghob. Literature T.G.Mattson, B.A.Sanders, B.L.Massingill: Patterns for Parallel Programming, Addison- Wesley,
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
Lecture 3 : Performance of Parallel Programs Courtesy : MIT Prof. Amarasinghe and Dr. Rabbah’s course note.
Programmability Hiroshi Nakashima Thomas Sterling.
Finding concurrency Jakub Yaghob. Finding concurrency design space Starting point for design of a parallel solution Analysis The patterns will help identify.
Exscale – when will it happen? William Kramer National Center for Supercomputing Applications.
DiRAC-3 – The future Jeremy Yates, STFC DiRAC HPC Facility.
Presented by PLASMA (Parallel Linear Algebra for Scalable Multicore Architectures) ‏ The Innovative Computing Laboratory University of Tennessee Knoxville.
Presented by PLASMA (Parallel Linear Algebra for Scalable Multicore Architectures) ‏ The Innovative Computing Laboratory University of Tennessee Knoxville.
Computing Systems: Next Call for Proposals Dr. Panagiotis Tsarchopoulos Computing Systems ICT Programme European Commission.
3/12/2013Computer Engg, IIT(BHU)1 INTRODUCTION-1.
Tackling I/O Issues 1 David Race 16 March 2010.
CERN VISIONS LEP  web LHC  grid-cloud HL-LHC/FCC  ?? Proposal: von-Neumann  NON-Neumann Table 1: Nick Tredennick’s Paradigm Classification Scheme Early.
Accelerating K-Means Clustering with Parallel Implementations and GPU Computing Janki Bhimani Miriam Leeser Ningfang Mi
Distributed Sequencing for Resource Sharing in Multi-Applicative Heterogeneous NoC Platforms 林鼎原 Department of Electrical Engineering National Cheng Kung.
Computer Science and Engineering Parallelizing Feature Mining Using FREERIDE Leonid Glimcher P. 1 ipdps’04 Scaling and Parallelizing a Scientific Feature.
Using Pattern-Models to Guide SSD Deployment for Big Data in HPC Systems Junjie Chen 1, Philip C. Roth 2, Yong Chen 1 1 Data-Intensive Scalable Computing.
Centre of Excellence in Physics at Extreme Scales Richard Kenway.
F1-17: Architecture Studies for New-Gen HPC Systems
Panel: Beyond Exascale Computing
Performance Assurance for Large Scale Big Data Systems
Productive Performance Tools for Heterogeneous Parallel Computing
Organizations Are Embracing New Opportunities
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Clouds , Grids and Clusters
Chandra S. Martha Min Lee 02/10/2016
Microarchitecture.
COMPUTATIONAL MODELS.
Seismic Hazard Analysis Using Distributed Workflows
For Massively Parallel Computation The Chaotic State of the Art
Extreme Big Data Examples
Parallel Programming By J. H. Wang May 2, 2017.
Unstructured Grids at Sandia National Labs
Parallel Programming in C with MPI and OpenMP
Model-Driven Analysis Frameworks for Embedded Systems
Adaptive Resource Allocation Technique for Exascale Systems
CPSC 531: System Modeling and Simulation
HPC User Forum 2012 Panel on Potential Disruptive Technologies Emerging Parallel Programming Approaches Guang R. Gao Founder ET International.
Power is Leading Design Constraint
Scalable Parallel Interoperable Data Analytics Library
Introduction to Operating Systems
Twister2: Design of a Big Data Toolkit
Dataflow Model of Computation (From Dataflow to Multithreading)
Panel on Research Challenges in Big Data
On the Role of Burst Buffers in Leadership-Class Storage Systems
Twister2 for BDEC2 Poznan, Poland Geoffrey Fox, May 15,
Presentation transcript:

Toward a Unified HPC and Big Data Runtime Joshua Suetterlein University of Delaware Joshua Landwehr, Joseph Manzano, Andres Marquez PNNL 11/18/2018

Why should we consider joining HPC and Big Data? Motivation Why should we consider joining HPC and Big Data? Boundaries are beginning to blur Next generation systems can cost Billions of $$$ in R&D and Millions to maintain Moreover technical challenges provide a unique opportunity HPC Big Data Computational science has become essential for engineering and science Reliably manages exponentially growing data Computational models enable the impossible (or impractical) Data analytics explores complex relationships among and growing data http://cacm.acm.org/magazines/2015/7/188732-exascale-computing-and-big-data/fulltext 11/18/2018

Motivation Cont. Known Exascale Challenges Energy efficiency Resilience Memory technology Interconnect technology Algorithms Scientific productivity http://ascr-discovery.science.doe.gov/2014/11/exascale-road-bumps/ Hidden Challenge: I/O vs Concurrency System 2009 2015 2024 Total Concurrency 225,000 O(million) O(billion) System Memory 0.3 PB 5 PB 10 PB I/O 0.2 TB 10 TB/s global PFS + 100 TB/s burst buffer 20 TB/s global PFS + 500 TB/s burst buffer Science at Extreme Scale: Architectural Challenges and Opportunities – Lucy Nowell April, 2014 11/18/2018

Motivation Cont. How will we perform science? Increased parallelism enables larger experiments Weak scaling – increase concurrency ≈ more data Current paradigm – offload results to another machine i.e. post-mortem analysis New paradigm – leverage In-situ/Big Data analytics then offload reduced results Another view: If a simulation is the source, we can leverage streaming analytics to hide IO latency Synergistic Challenges in Data-Intensive Science and Exascale Computing. March, 2013 11/18/2018

Fine-Grain How is computational science traditionally performed? MPI + OpenMP dominate scientific code Architectural trends previously amenable to OpenMP and MPI are changing Unprecedented concurrency Highly coupled memory and processing elements Deepening memory hierarchies Increasingly heterogeneity Strict energy and power budgets A dataflow inspired, asynchronous, event driven, fine-grain execution model appears flexible enough to express and exploit sufficient parallelism to utilize new architectures Gao. 7/2002 Fran Allen’s Retirement Workshop 11/18/2018

with Big Data Extensions Fine-Grain Cont. Open Community Runtime (OCR) is a codelet execution model developed under UHPC and X-Stack initiatives Exploits asynchronous, event driven, fine-grain execution Can OCR be augmented with streaming Big Data techniques to reduce IO pressures and hide latency? Application Fine-grain runtime with Big Data Extensions RDMA Scheduler (SLURM) File System (Lustre) Compute Nodes 11/18/2018

Opportunities Fine-grain runtimes have the potential of providing more than weak scaling For Exascale, system memory is getting bigger (~10PB) Big Data trends provide a new abstractions for analytics In-situ has not been very flexible Events, memory abstractions, and compression are already being explored 11/18/2018

Challenges Programming model disconnect Load balancing How can we effectively map simulated data generated by fine-grain tasks into partitioned key/value Load balancing Transparent parallel execution of a partition of data Priorities: Simulation vs Analysis This is similar to streaming back-pressure except we control the source… How does this effect the critical path? 11/18/2018

Questions Thank You! 11/18/2018