Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb. 2015.

Slides:



Advertisements
Similar presentations
1 Copyright © 2012 Oracle and/or its affiliates. All rights reserved. Convergence of HPC, Databases, and Analytics Tirthankar Lahiri Senior Director, Oracle.
Advertisements

Issues of HPC software From the experience of TH-1A Lu Yutong NUDT.
The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
SAN DIEGO SUPERCOMPUTER CENTER Niches, Long Tails, and Condos Effectively Supporting Modest-Scale HPC Users 21st High Performance Computing Symposia (HPC'13)
PRAKTICKÝ ÚVOD DO SUPERPOČÍTAČE ANSELM Infrastruktura, přístup a podpora uživatelů David Hrbáč
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
WEST VIRGINIA UNIVERSITY HPC and Scientific Computing AN OVERVIEW OF HIGH PERFORMANCE COMPUTING RESOURCES AT WVU.
Information Technology Center Introduction to High Performance Computing at KFUPM.
JLab Status & 2016 Planning April 2015 All Hands Meeting Chip Watson Jefferson Lab Outline Operations Status FY15 File System Upgrade 2016 Planning for.
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME SESAME – LinkSCEEM.
ASKAP Central Processor: Design and Implementation Calibration and Imaging Workshop 2014 ASTRONOMY AND SPACE SCIENCE Ben Humphreys | ASKAP Software and.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
UNCLASSIFIED: LA-UR Data Infrastructure for Massive Scientific Visualization and Analysis James Ahrens & Christopher Mitchell Los Alamos National.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Dell IT Innovation Labs in the Cloud “The power to do more!” Andrew Underwood – Manager, HPC & Research Computing APJ Solutions Engineering Team.
Illinois Campus Cluster Program User Forum October 24, 2012 Illini Union Room 210 2:00PM – 3:30PM.
The University of Texas Research Data Repository : “Corral” A Geographically Replicated Repository for Research Data Chris Jordan.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
NSF Vision and Strategy for Advanced Computational Infrastructure Vision: NSF Leadership in creating and deploying a comprehensive portfolio…to facilitate.
DNA Subway Green Line Overview. Growth of Sequence Read Archive (SRA) 2.2 Quadrillion bases Log Scale!
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
Project Overview:. Longhorn Project Overview Project Program: –NSF XD Vis Purpose: –Provide remote interactive visualization and data analysis services.
GPU Programming with CUDA – Accelerated Architectures Mike Griffiths
High Performance Computing G Burton – ICG – Oct12 – v1.1 1.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
1 Advanced Storage Technologies for High Performance Computing Sorin, Faibish EMC NAS Senior Technologist IDC HPC User Forum, April 14-16, Norfolk, VA.
LARGE SCALE DEPLOYMENT OF DAP AND DTS Rob Kooper Jay Alemeda Volodymyr Kindratenko.
Introduction to HPC resources for BCB 660 Nirav Merchant
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Big Red II & Supporting Infrastructure Craig A. Stewart, Matthew R. Link, David Y Hancock Presented at IUPUI Faculty Council Information Technology Subcommittee.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
The Cray XC30 “Darter” System Daniel Lucio. The Darter Supercomputer.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
SAN DIEGO SUPERCOMPUTER CENTER SDSC's Data Oasis Balanced performance and cost-effective Lustre file systems. Lustre User Group 2013 (LUG13) Rick Wagner.
A Framework for Visualizing Science at the Petascale and Beyond Kelly Gaither Research Scientist Associate Director, Data and Information Analysis Texas.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
National Center for Supercomputing Applications University of Illinois at Urbana–Champaign Visualization Support for XSEDE and Blue Waters DOE Graphics.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
Accelerating High Performance Cluster Computing Through the Reduction of File System Latency David Fellinger Chief Scientist, DDN Storage ©2015 Dartadirect.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Remote & Collaborative Visualization. TACC Remote Visualization Systems Longhorn – Dell XD Visualization Cluster –256 nodes, each with 48 GB (or 144 GB)
Parallel IO for Cluster Computing Tran, Van Hoai.
Tackling I/O Issues 1 David Race 16 March 2010.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Petascale Computing Resource Allocations PRAC – NSF Ed Walker, NSF CISE/ACI March 3,
Scaling up R computation with high performance computing resources.
The Evolution of the Italian HPC Infrastructure Carlo Cavazzoni CINECA – Supercomputing Application & Innovation 31 Marzo 2015.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Slide 1 User-Centric Workload Analytics: Towards Better Cluster Management Saurabh Bagchi Purdue University Joint work with: Subrata Mitra, Suhas Javagal,
Slide 1 Cluster Workload Analytics Revisited Saurabh Bagchi Purdue University Joint work with: Subrata Mitra, Suhas Javagal, Stephen Harrell (Purdue),
Compute and Storage For the Farm at Jlab
A Brief Introduction to NERSC Resources and Allocations
Buying into “Summit” under the “Condo” model
Early Results of Deep Learning on the Stampede2 Supercomputer
Low-Cost High-Performance Computing Via Consumer GPUs
Jay Boisseau, Director Texas Advanced Computing Center
Scaling Spark on HPC Systems
Scott Michael Indiana University July 6, 2017
USF Health Informatics Institute (HII)
Early Results of Deep Learning on the Stampede2 Supercomputer
Presentation transcript:

Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb. 2015

Agenda 8:30-9:00 Welcome and introduction to TACC resources. 9:00–9:30 Getting started with running R at TACC. 9:30–10:00 Practice and coffee break. 10:00-11:00 R basics 11:00-11:30 Data analysis support in R 11:30-1:00 Lunch break 1:00-1:30 Scaling up R computations 1:30-2:00 A walkthrough with parallel package in R 2:00-3:00 hands on lab session 3:00-4:00 Understand the performance of R program

Introduction to TACC Resource

About TACC TACC is a Research Division at the University of Texas at Austin –Origins go back to 1960s Cray CDC 6600 support –TACC started in 2001 to support research beyond UT needs TACC is a service provider for XSEDE on several key systems –Currently providing between 80 to 90% of HPC cycles in XSEDE –Not limited to supporting NSF research TACC is also supported by partnering with UT Austin, UT System, Industrial Partners, Multi-institutional research grants, and donations TACC is 110+ people (40+ PhDs) bringing enabling technologies and techniques to drive digital research –Many collaborative research projects and mission specific proposals to support open research –Consulting to bring TACC expertise to other communities

Stampede Base Cluster (Dell/Intel/Mellanox): –6,400 nodes –Intel Sandy Bridge processors –Dell dual-socket nodes w/32GB RAM (2GB/core) –56 Gb/s Mellanox FDR InfiniBand interconnect –More than 100,000 cores, 2.2 PF peak performance Max Total Concurrency: –exceeds 500,000 cores –1.8M threads –#7 in HPC top % allocated through XSEDE

Additional Features of Stampede 6800 Intel Xeon Phi “MIC” Many Integrated Core processors –Special release of “Knight’s Corner” (61 cores) –10+ PF peak performance Stampede includes 16 1TB Sandy Bridge shared memory nodes with dual GPUs 128 of the compute nodes are also equipped with NVIDIA Kepler K20 GPUs Storage subsystem driven by Dell storage nodes: –Aggregate Bandwidth greater than 150GB/s –More than 14PB of capacity –Similar partitioning of disk space into multiple Lustre filesystems as previous TACC systems ($HOME, $WORK and $SCRATCH)

What does this mean? Faster processors More memory per node Starting hundreds of analysis jobs in batch. Access to latest “massive parallel” hardware –Intel Xeon Phi –GPGPU

Automatic offloading with latest hardware R is originally designed as for single thread execution. –Slow performance –Not scalable with large data R can be built and linked to library utilizes latest multiple core technology for automatic parallel execution for some operations, most commonly, linear algebra related computations.

Getting more from R Optimizing R performance on Stampede –Intel compiler vs. gcc was a factor of 2 improvement –MKL significantly improved performance –Some Xeon Phi performance enhancement too –Supporting common parallel packages

Maverick Hardware 132 Node dual core Ivy Bridge based cluster –Each node has NVIDIA Kepler K40 GPU –128 GB of memory –FDR Interconnect –Shares Work file system with Stampede (26 PB unformatted) –Users get 1 TB of Work to start Intended for real time analysis TACC system, 50% provided to XSEDE in kind, 50% discretionary

Visualization and Analysis Portal

R and Python Can launch RStudio Server and iPython Notebook –Introducing capabilities, best practices, and forms of parallelism to users –Simplifying UI with web interface –Improving visualization capabilities with Shiny package and GoogleVis

Hadoop Cluster: Rustler A Hadoop cluster with 64 Hadoop Data Nodes –2 x 10 core Ivy Bridge processors –128 GB memory –16x1TB disks (1 PB usable disk, 333 TB replicated) Login node, 2 Name nodes, 1 Web Proxy node 10 Gb/s Ethernet network with 40 Gb/s connectivity to TACC backbone In early user period today A pure TACC resource (All discretionary allocations)

Wrangler Three primary subsystems: –A 10PB disk storage system –Lustre based – (2 R720 dual E MD servers, 45 C8000 OSF servers with 6 TB drives) –An embedded analytics capability of several thousand cores. –96 Dell R620 Haswell E v3 nodes with dual IB FDR/40 Gb/s Ethernet –A high speed global object store 500 TB usable Flash via PCI to all 96 analytics nodes 1TB/s IO rate &250M+ IOPS

Data Intensive Computing Support at TACC Data Management and Collection group –Providing data storage service Files, databases, irods, –Collection management and curation Data Mining and Statistics group –Collaborating with users to develop and implement scalable algorithmic solution. –In addition to general data mining and analysis method, also expertise in R, Hadoop and visual analytics.. We are here to help: