Presentation is loading. Please wait.

Presentation is loading. Please wait.

Digital Science Center

Similar presentations


Presentation on theme: "Digital Science Center"— Presentation transcript:

1 Digital Science Center
Geoffrey C. Fox, David Crandall, Judy Qiu, Gregor von Laszewski, Fugang Wang, Badi' Abdul-Wahid Saliya Ekanayake, Supun Kamburugamuva, Jerome Mitchell, Bingjing Zhang School of Informatics and Computing, Indiana University Digital Science Center Research Areas Green implies HPC Integration Scientific Impact Metrics We developed a software framework and process to evaluate scientific impact for XSEDE. We calculate and track various Scientific Impact Metrics of XSEDE., BlueWaters, and NCAR. In recently conducted peer comparison study we showed how well XSEDE and its support services (ECSS) perform judging by citations received. During this process we retrieved and processed millions of data entries from multiple sources in various formats to obtain the result. Digital Science Center Facilities RaPyDLI Deep Learning Environment SPIDAL Scalable Data Analytics Library and applications including Ice Sheet MIDAS Big Data Software Big Data Ogres Classification and Benchmarks CloudIOT Internet of Things Environment Cloudmesh Cloud and Bare metal Automation XSEDE TAS Monitoring citations and system metrics Data Science Education with MOOC’s # Publications Rank - Average Rank - Median # Citation - Average # Citation - Median XSEDE 2349 61 65 26 11 Peers 168422 49 48 13 5 Ice Layer Detection Algorithm  Data analytics for IoT devices in Cloud Parallel Sparse LDA  High Performance Data Analytics with Java + MPI on Multicore HPC Clusters We developed a framework to bring data from IoT devices to a cloud environment for real time data analysis. The framework consists of; Data collection nodes near the devices, Publish-subscribe brokers to bring data to cloud and Apache Storm coupled with other batch processing engines for data processing in cloud. Our data pipe line is Robot → Gateway → Message Brokers → Apache Storm. Original LDA (orange) compared to LDA exploiting sparseness (blue) Note data analytics making use of Infiniband (i.e. limited by communication!) Java code running under Harp – Hadoop plus HPC plugin Corpus: 3,775,554 Wikipedia documents, Vocabulary: 1 million words; Topics: 10k topics; BR II is Big Red II supercomputer with Cray Gemini interconnect Juliet is Haswell Cluster with Intel (switch) and Mellanox (node) infiniband (not optimized) Harp LDA on BR2 The polar science community has built radars capable of surveying the polar ice sheets, and as a result, have collected terabytes of data and is increasing its repository each year as signal processing techniques improve and the cost of hard drives decrease enabling a new-generation of high resolution ice thickness and accumulation maps. Manually extracting layers from an enormous corpus of ice thickness and accumulation data is time-consuming and requires sparse hand-selection, so developing image processing techniques to automatically aid in the discovery of knowledge is of high importance. We find it is challenging to achieve high performance in HPC clusters for big data problems. We approach this with Java and MPI, but improves further using Java memory maps to and off heap data structures. We achieve zero intra-node messaging, zero GC, and minimal memory footprint. We present performance results of running it on a latest Intel Haswell HPC cluster consisting 3456 cores total. Simultaneous Localization and Mapping(SLAM) is an example application built on top of our framework, where we exploit parallel data processing to speedup the expensive SLAM computation. Robot Map of the Environment Harp LDA on Juliet (36 core nodes)

2 Digital Science Center
Geoffrey C. Fox, David Crandall, Judy Qiu, Gregor von Laszewski, Fugang Wang, Badi' Abdul-Wahid, Saliya Ekanayake, Supun Kamburugamuva, Jerome Mitchell, Bingjing Zhang School of Informatics and Computing, Indiana University DSC Computing Systems FutureSystems Use of FutureSystems India Cluster Rapid Prototyping HPC Environment for Deep Learning Infrastructure HPC and cloud mixed environment Had openstack in service since Cactus release back to 2011 Currently operating two openstack clouds: juno and kilo, with ~60 compute nodes total. Usage Data of OpenStack 1,060,442 hours of wall time (between 01/01/ /31/2015) 14,318 VM instances launched total  (between 01/01/ /31/2015) Average 63 VMs launched per day 70% of tiny or small server size and 30% of medium or (x)large server size were used Courses Total Courses to date: 75 course projects from 31 distinct institutions since 2010 SP15 v594 (online)Topic: cloud computing, big data Audience: #38 globally distributed GE employees Projects: deployments of big data platforms and analysis of data sets FA15, FA14 (online) [in progress] Topic: big data Audience: #139 students (on site & remote) Projects: analysis of sports, Amazon movie reviews, stock market, twitter data Undergraduate Student Research Developing Cloudmesh for cloud environment management REU and on-site students Just installed 128 node Haswell based system (Juliet) 128 GB memory per node Substantial conventional disk per node (8TB) plus PCI based SSD Infiniband with SR-IOV 24 and 36 core nodes (3456 total cores) Working with SDSC on NSF XSEDE Comet System (Haswell 47,776 cores) Older machines India (128 nodes, 1024 cores), Bravo (16 nodes, 128 cores), Delta(16 nodes, 192 cores), Echo(16 nodes, 192 cores), Tempest (32 nodes, 768 cores) with large memory, large disk and GPU Optimized for Cloud research and Large scale Data analytics exploring storage models, algorithms Build technology to support high performance virtual clusters We are developing an on-ramp to deep learning that utilizes HPC and GPU resources. It serves as an aggregate for modules related to deep learning. We address a multitude of issues including deployment, access, and integration into HPC environments. Simple interfaces are provided that allow easy reusability. We are working with our partners on a performance optimized convolution kernel that will be able to utilize state of the art GPUs including AMD. Data management will be available as part of deep learning workfflows. JULIET INDIA TEMPEST BRAVO-DELTA-ECHO Cloudmesh for Managing Virtual Clusters Virtual Clusters with SDSC Comet  Threads v. Processes on 24 core Juliet Nodes for 48 and 96 nodes Threads v. Processes on 24 core Juliet Nodes for 24, 48 and 96 nodes Comet, is a new petascale supercomputer designed to transform advanced scientific computing by expanding access and capacity among traditional as well as non-traditional research domains. Comet will be capable of an overall peak performance of nearly two petaflops, or two quadrillion operations per second. We are working together with SDSC to deliver an easy to use service that allows users to request virtual clusters that are close to hardware. Advanced users will be able to request such clusters and allow the users to manage them and deploy their own software stacks on them. Comet is the first virtualized HPC cluster, and delivers a significantly increased level of computing capacity and customizability to support data-enabled science and engineering at the campus, regional, and national levels, and in turn support the entire science and engineering enterprise, including education as well as research. IU is currently working on delivering easy to use client interfaces and leverages experiences from our cloudmesh software. Cloudmesh is an important component designed to deliver a software-defined system – encompassing virtualized and bare-metal infrastructure, networks, application, systems and platform software – with a unifying goal of providing Cloud Testbeds as a Service (CTaaS). Cloudmesh federates a number of resources from academia and industry. This includes existing FutureSystems, Amazon Web Services, Azure, and HP Cloud.


Download ppt "Digital Science Center"

Similar presentations


Ads by Google