Department of Intelligent Systems Engineering

Slides:



Advertisements
Similar presentations
HPC-ABDS: The Case for an Integrating Apache Big Data Stack with HPC
Advertisements

Iterative computation is a kernel function to many data mining and data analysis algorithms. Missing in current MapReduce frameworks is collective communication,
Data Science at Digital Science October Geoffrey Fox Judy Qiu
Harp: Collective Communication on Hadoop Bingjing Zhang, Yang Ruan, Judy Qiu.
SALSASALSA Large-Scale Data Analysis Applications Computer Vision Complex Networks Bioinformatics Deep Learning Data analysis plays an important role in.
Indiana University Faculty Geoffrey Fox, David Crandall, Judy Qiu, Gregor von Laszewski Data Science at Digital Science Center.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
1 Panel on Merge or Split: Mutual Influence between Big Data and HPC Techniques IEEE International Workshop on High-Performance Big Data Computing In conjunction.
Geoffrey Fox Panel Talk: February
Digital Science Center
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Next Generation Grid: Integrating Parallel and Distributed Computing Runtimes for an HPC Enhanced Cloud and Fog Spanning IoT Big Data and Big Simulations.
SPIDAL Analytics Performance February 2017
Digital Science Center II
Department of Intelligent Systems Engineering
Next Generation IoT and Data-based Grid
Status and Challenges: January 2017
Spark Presentation.
Big Data and High-Performance Technologies for Natural Computation
NSF start October 1, 2014 Datanet: CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science Indiana University.
Interactive Website (
Research in Digital Science Center
Big Data Processing Issues taking care of Application Requirements, Hardware, HPC, Grid (distributed), Edge and Cloud Computing Geoffrey Fox, November.
Next Generation Grid: Integrating Parallel and Distributed Computing Runtimes from Cloud to Edge Applications The 15th IEEE International Symposium on.
Some Remarks for Cloud Forward Internet2 Workshop
NSF : CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science PI: Geoffrey C. Fox Software: MIDAS HPC-ABDS.
Department of Intelligent Systems Engineering
Digital Science Center I
HPSA18: Logistics 7:00 am – 8:00 am Breakfast
Twister2: A High-Performance Big Data Programming Environment
I590 Data Science Curriculum August
Versatile HPC: Comet Virtual Clusters for the Long Tail of Science SC17 Denver Colorado Comet Virtualization Team: Trevor Cooper, Dmitry Mishin, Christopher.
High Performance Big Data Computing in the Digital Science Center
Convergence of HPC and Clouds for Large-Scale Data enabled Science
Research in Intelligent Systems Engineering
Data Science Curriculum March
HPC-enhanced IoT and Data-based Grid
Department of Intelligent Systems Engineering
Tutorial Overview February 2017
Department of Intelligent Systems Engineering
AI First High Performance Big Data Computing for Industry 4.0
Data Science for Life Sciences Research & the Public Good
13th Cloud Control Workshop, June 13-15, 2018
A Tale of Two Convergences: Applications and Computing Platforms
Martin Swany Gregor von Laszewski Thomas Sterling Clint Whaley
Research in Digital Science Center
Scalable Parallel Interoperable Data Analytics Library
Cloud DIKW based on HPC-ABDS to integrate streaming and batch Big Data
HPC Cloud and Big Data Testbed
High Performance Big Data Computing
10th IEEE/ACM International Conference on Utility and Cloud Computing
Twister2: Design and initial implementation of a Big Data Toolkit
Indiana University, Bloomington
Twister2: Design of a Big Data Toolkit
Department of Intelligent Systems Engineering
2 Programming Environment for Global AI and Modeling Supercomputer GAIMSC 2/19/2019.
Introduction to Twister2 for Tutorial
$1M a year for 5 years; 7 institutions Active:
PHI Research in Digital Science Center
Panel on Research Challenges in Big Data
Assoc. Prof. Marc FRÎNCU, PhD. Habil.
Big Data, Simulations and HPC Convergence
Motivation Contemporary big data tools such as MapReduce and graph processing tools have fixed data abstraction and support a limited set of communication.
High-Performance Big Data Computing
Big Data and High-Performance Technologies for Natural Computation
Research in Digital Science Center
Geoffrey Fox High-Performance Big Data Computing: International, National, and Local initiatives COLLABORATORS China and IU: Fudan University, SICE, OVPR.
Research in Digital Science Center
Convergence of Big Data and Extreme Computing
Twister2 for BDEC2 Poznan, Poland Geoffrey Fox, May 15,
Presentation transcript:

Department of Intelligent Systems Engineering Engineering Cyberinfrastructure in Intelligent Systems Engineering at Indiana University Geoffrey Fox, Judy Qiu, Martin Swany, Thomas Sterling, Gregor von Laszewski November 20, 2017 Department of Intelligent Systems Engineering http://www.dsc.soic.indiana.edu/ `

Martin Swany: Photon Lightweight Messaging Photon provides consistent remote direct memory access (RDMA) semantics over varied interconnect technologies such as InfiniBand and Cray’s Aries and Gemini fabrics. Goal: minimize latency and maximize throughput for high- performance applications and runtime systems that can benefit from distributed, direct memory operations over a network. Memory management and asynchronous progress are exposed at a fine granularity, decoupling data transfers from the notification path. Photon supports active messages.

Swany: FPGA Machines Node of 64 node ISE FPGA Lab Machine Proteus with Xilinx boards aimed at computing and networking Currently have an 8 node prototype shown on right 4-core ARM + FPGA

Thomas Sterling CREST-II R&D Program Next generation exascale (supercomputing) hardware and software Strategic Focus Projects Tactical Focus Projects HPX+ Parallel runtime Software Software Runtime System For conventional and CCA targets Reduction to practice for ParalleX Fault tolerance, energy, real-time Advanced introspective scheduling CCA Dataflow Hardware Continuum Computer Architecture CCA Fonton architecture Simultac architecture Name space translation Routing of parcels System engineering Packaging, power, cooling ParalleX Programming Model Execution model Formal specification & testing Dynamic adaptive, introspective Fault tolerance, energy, real-time Using DIET high speed networking hardware Programming Interface Intermediate representation Compiler for runtime system Graph algorithms/applications Meta-data parallelism discovery Concurrent edge spanning

Ogres Big Data Application Feature Analysis Fox/ Qiu/Crandall NSF 1443054: CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science Ogres Big Data Application Feature Analysis Midas Middleware: HPC-ABDS and HPC-FaaS Software Harp and Twister2 Building Blocks SPIDAL Data Analytics Library Software: MIDAS HPC-ABDS Large Interdisciplinary Project – 7 institutions Cyberinfrastructure – Algorithms - Applications

Qiu/Fox Core SPIDAL Parallel HPC Library with Collective Used QR Decomposition (QR) Reduce, Broadcast DAAL Neural Network AllReduce DAAL Covariance AllReduce DAAL Low Order Moments Reduce DAAL Naive Bayes Reduce DAAL Linear Regression Reduce DAAL Ridge Regression Reduce DAAL Multi-class Logistic Regression Regroup, Rotate, AllGather Random Forest AllReduce Principal Component Analysis (PCA) AllReduce DAAL DA-MDS Rotate, AllReduce, Broadcast Directed Force Dimension Reduction AllGather, Allreduce Irregular DAVS Clustering Partial Rotate, AllReduce, Broadcast DA Semimetric Clustering Rotate, AllReduce, Broadcast K-means AllReduce, Broadcast, AllGather DAAL SVM AllReduce, AllGather SubGraph Mining AllGather, AllReduce Latent Dirichlet Allocation Rotate, AllReduce Matrix Factorization (SGD) Rotate DAAL Recommender System (ALS) Rotate DAAL Singular Value Decomposition (SVD) AllGather DAAL DAAL implies integrated with Intel DAAL Optimized Data Analytics Library (Runs on KNL!) 9/18/2018

Judy Qiu: Harp-DAAL for high performance machine learning HPC-ABDS is Cloud-HPC interoperable software with the performance of HPC (High Performance Computing) and the rich functionality of the commodity Apache Big Data Stack. This concept is illustrated by Harp-DAAL. High Level Usability: Python Interface, well documented and packaged modules Middle Level Data-Centric Abstractions: Computation Model and optimized communication patterns Low Level optimized for Performance: HPC kernels Intel® DAAL and advanced hardware platforms such as Xeon and Xeon Phi

Qiu MIDAS run time software for Harp broadcast reduce allreduce allgather regroup push & pull rotate Map Collective Run time merges MapReduce and HPC

Harp v. Spark Harp v. Torch Harp v. MPI Datasets: 5 million points, 10 thousand centroids, 10 feature dimensions 10 to 20 nodes of Intel KNL7250 processors Harp-DAAL has 15x speedups over Spark MLlib Datasets: 500K or 1 million data points of feature dimension 300 Running on single KNL 7250 (Harp-DAAL) vs. single K80 GPU (PyTorch) Harp-DAAL achieves 3x to 6x speedups Datasets: Twitter with 44 million vertices, 2 billion edges, subgraph templates of 10 to 12 vertices 25 nodes of Intel Xeon E5 2670 Harp-DAAL has 2x to 5x speedups over state-of-the-art MPI-Fascia solution

WRF Weather Modeling Virtual Cluster Performance 96-core (4-node) virtual cluster Nearest-neighbor communication Test Case: 3hr Forecast, 2.5km resolution of Continental US (CONUS). Scalable algorithms 2% slower w/ SR-IOV vs native IB. Gregor von Laszewski Virtual Cluster Goal: Provide near bare metal HPC performance and management experience for groups that can manage their own clusters. Virtual Cluster Enabling Technologies KVM—Lets us run virtual machines (all processor features) SR-IOV—Makes MPI go fast on VMs Rocks—Systems management ZFS—Disk image management VLANs—Isolate virtual cluster management network pkeys—Isolate virtual cluster IB network Nucleus —Coordination (scheduling, provisioning, status, etc.) Client – Cloudmesh control of DevOps WRF Weather Modeling

Geoffrey Fox Twister2: “Next Generation Grid - Edge – HPC Cloud” The original 2010 Twister paper was a particular approach to MapCollective iterative processing for machine learning improving Hadoop (Harp also builds on Twister) Twister2 re-engineers current Apache Big Data and HPC software systems as a toolkit Support a serverless (cloud-native) dataflow event-driven HPC-FaaS (microservice) framework running across application and geographic domains. Support all types of Data analysis from Global Machine Learning to Edge computing Build on Cloud best practice but use HPC wherever possible to get high performance Smoothly support current paradigms Hadoop, Spark, Flink, Heron, MPI, DARMA … Includes IoTCloud Cloud computing Robot/IoT control software Use interoperable common abstractions but multiple polymorphic implementations. Focus on Runtime but this implies HPC-FaaS programming and execution model This defines a next generation Grid based on data and edge devices – not computing as in old Grid See paper http://dsc.soic.indiana.edu/publications/twister2_design_big_data_toolkit.pdf

Intelligent System Engineering Computing Infrastructure For research and education ISE arranged as a collection of centers and laboratories 64-node FPGA+ARM + four 10G Ethernet Cluster Proteus 16 K80 GPU + 16 P100 Volta Deep Learning Cluster Romeo (Minje Kim, David Crandall) 128 node Haswell + Infiniband Machine Learning Testbed Juliet 64 node Intel Knights Landing + Omnipath Machine Learning Testbed Tango 24 node Xeon Platinum + Infiniband Cloud Computing Cluster Victor Plus University has substantial supercomputing resources Big Red 2