1 Multicore and Cloud Futures CCGSC September 15 2008 Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University

Slides:



Advertisements
Similar presentations
Introduction to Grid Application On-Boarding Nick Werstiuk
Advertisements

A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
SALSA HPC Group School of Informatics and Computing Indiana University.
Android architecture overview
Clouds from FutureGrid’s Perspective April Geoffrey Fox Director, Digital Science Center, Pervasive.
1 Overview of Cyberinfrastructure and the Breadth of Its Application Geoffrey Fox Computer Science, Informatics, Physics Chair Informatics Department Director.
Building Science Gateways Marlon Pierce Community Grids Laboratory Indiana University.
1 Clouds and Sensor Grids CTS2009 Conference May Alex Ho Anabas Inc. Geoffrey Fox Computer Science, Informatics, Physics Chair Informatics Department.
University of Illinois Role of Mashups, Cloud Computing, and Parallelism for Visual Analytics Loretta Auvil.
1 Earthquake Polar and Sensor Grids Community Grids Laboratory November Geoffrey Fox Community Grids Laboratory, School of informatics Indiana.
Student Visits August Geoffrey Fox
Parallel Data Analysis from Multicore to Cloudy Grids Indiana University Geoffrey Fox, Xiaohong Qiu, Scott Beason, Seung-Hee.
EXTENDING SCIENTIFIC WORKFLOW SYSTEMS TO SUPPORT MAPREDUCE BASED APPLICATIONS IN THE CLOUD Shashank Gugnani Tamas Kiss.
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
HPC-ABDS: The Case for an Integrating Apache Big Data Stack with HPC
SALSASALSA Programming Abstractions for Multicore Clouds eScience 2008 Conference Workshop on Abstractions for Distributed Applications and Systems December.
1 Challenges Facing Modeling and Simulation in HPC Environments Panel remarks ECMS Multiconference HPCS 2008 Nicosia Cyprus June Geoffrey Fox Community.
Big Data and Clouds: Challenges and Opportunities NIST January Geoffrey Fox
1 1 Hybrid Cloud Solutions (Private with Public Burst) Accelerate and Orchestrate Enterprise Applications.
Science in Clouds SALSA Team salsaweb/salsa Community Grids Laboratory, Digital Science Center Pervasive Technology Institute Indiana University.
PolarGrid Geoffrey Fox (PI) Indiana University Associate Dean for Graduate Studies and Research, School of Informatics and Computing, Indiana University.
DISTRIBUTED COMPUTING
OpenQuake Infomall ACES Meeting Maui May Geoffrey Fox
IU QuakeSim/E-DECIDER Effort. QuakeSim Accomplishments (1) Deployed, improved many QuakeSim gadgets for standalone integration into QuakeSim.org – Disloc,
1 Performance of a Multi-Paradigm Messaging Runtime on Multicore Systems Poster at Grid 2007 Omni Austin Downtown Hotel Austin Texas September
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
Parallel Applications And Tools For Cloud Computing Environments Azure MapReduce Large-scale PageRank with Twister Twister BLAST Thilina Gunarathne, Stephen.
GEM Portal and SERVOGrid for Earthquake Science PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics, Physics.
SALSA HPC Group School of Informatics and Computing Indiana University.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Service - Oriented Middleware for Distributed Data Mining on the Grid ,劉妘鑏 Antonio C., Domenico T., and Paolo T. Journal of Parallel and Distributed.
HPC User Forum Back End Compiler Panel SiCortex Perspective Kevin Harris Compiler Manager April 2009.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
1 Performance Measurements of CCR and MPI on Multicore Systems Expanded from a Poster at Grid 2007 Austin Texas September Xiaohong Qiu Research.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
QuakeSim Project: Portals and Web Services for Geophysics Marlon Pierce Indiana University
ISERVOGrid Architecture Working Group Brisbane Australia June Geoffrey Fox Community Grids Lab Indiana University
QuakeSim Project: Portals and Web Services for Geo-Sciences Marlon Pierce Indiana University
SALSASALSASALSASALSA Clouds Ball Aerospace March Geoffrey Fox
Message Management April Geoffrey Fox Computer Science, Informatics, Physics Pervasive Technology Laboratories Indiana University Bloomington IN.
SALSASALSASALSASALSA Cloud Panel Session CloudCom 2009 Beijing Jiaotong University Beijing December Geoffrey Fox
Virtual Appliances CTS Conference 2011 Philadelphia May Geoffrey Fox
1 Multicore for Science Multicore Panel at eScience 2008 December Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University.
QuakeSim Project: Portals and Web Services for Geo-Sciences Marlon Pierce Indiana University
Looking at Use Case 19, 20 Genomics 1st JTC 1 SGBD Meeting SDSC San Diego March Judy Qiu Shantenu Jha (Rutgers) Geoffrey Fox
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
SALSASALSASALSASALSA Digital Science Center February 12, 2010, Bloomington Geoffrey Fox Judy Qiu
Building Science Gateways Marlon Pierce Community Grids Laboratory Indiana University.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
HPC in the Cloud – Clearing the Mist or Lost in the Fog Panel at SC11 Seattle November Geoffrey Fox
1 Cloud Systems Panel at HPDC Boston June Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University
NGS computation services: APIs and.
Directions in eScience Interoperability and Science Clouds June Interoperability in Action – Standards Implementation.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
INFN OCCI implementation on Grid Infrastructure Michele Orrù INFN-CNAF OGF27, 13/10/ M.Orrù (INFN-CNAF) INFN OCCI implementation on Grid Infrastructure.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Community Grids Laboratory
For Massively Parallel Computation The Chaotic State of the Art
Our Objectives Explore the applicability of Microsoft technologies to real world scientific domains with a focus on data intensive applications Expect.
I590 Data Science Curriculum August
Biology MDS and Clustering Results
GCC2008 (Global Clouds and Cores 2008) October Geoffrey Fox
Summary Background Introduction in algorithms and applications
Clouds from FutureGrid’s Perspective
HPML Conference, Lyon, Sept 2018
3 Questions for Cluster and Grid Use
Technology Futures and Lots of Sensor Grids
CReSIS Cyberinfrastructure
Technology Futures and Lots of Sensor Grids
Clouds and Grids Multicore and all that
Presentation transcript:

1 Multicore and Cloud Futures CCGSC September Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University

2 Gartner 2008 Technology Hype Curve Clouds, Microblogs and Green IT appear Basic Web Services, Wikis and SOA becoming mainstream MPI way out on plateau? Grids?

C YBERINFRASTRUCTURE C ENTER FOR P OLAR S CIENCE (CICPS) 3 Gartner 2006 Technology Hype Curve Grids did exist

Grids become Clouds Grids solve problem of too little computing: We need to harness all the world’s computers to do Science Clouds solve the problem of too much computing: with multicore we have so much power that we need to make usage much easier Key technology: Virtual Machines (dynamic deployment) enable more dynamic flexible environments Is Virtual Cluster or Virtual Machine correct primitive? Data Grids seem fine as data naturally distributed GGF/EGA false assumption: Web 2.0 not Enterprise defined commercial software stack Some Web 2.0 applications (MapReduce) not so different from data-deluged eScience Citizen Science requires light weight friendly Cyberinfrastructure 4

MPI on Nimbus for clustering Note fluctuations in runtime but performance OK for large enough problems 8 Nodes 5

Plans for QuakeSpace QuakeSim supports Earthquake Scientists who want some features of their kid’s (under 40) world Rebuild QuakeSim using Web 2.0 and Cloud Technology Applications, Sensors, Data Repositories as Services Computing via Clouds Portals as Gadgets Metadata by tagging Data sharing as in YouTube Alerts by RSS Virtual Organizations via Social Networking Workflow by Mashups Performance by multicore Interfaces via iPhone, Android etc. 6

Enterprise ApproachWeb 2.0 Approach JSR 168 PortletsGadgets, Widgets Server-side integration and processing AJAX, client-side integration and processing, JavaScript SOAPRSS, Atom, JSON WSDLREST (GET, PUT, DELETE, POST) Portlet ContainersOpen Social Containers (Orkut, LinkedIn, Shindig); Facebook; StartPages User Centric GatewaysSocial Networking Portals Workflow managers (Taverna, Kepler, etc) Mash-ups Grid computing: Globus, Condor, etcCloud computing: Amazon WS Suite, Xen Virtualization, still Condor!

Different Programming Models (Web) services, "farm" computations, Workflow (including AVS, HeNCE from past), Mashups, MPI, MapReduce run functionally or data decomposed execution units with a wide variety of front ends Front-end: Language+communication library, Scripting, Visual, Functional, XML, PGAS, HPCS Parallel Languages, Templates, OpenMP Synchronize/Communicate with some variant of messaging (zero size for locks) with performance, flexibility, fault-tolerance, dynamism trade-offs Synchronization: Locks Threads Processes CCR CCI SOAP REST MPI Hadoop; not much difference for user? 9

MPI becomes Ghetto MPI Multicore best practice not messaging will drive synchronization/communication primitives Party Line Programming Model: Workflow (parallel-- distributed) controlling optimized library calls Core parallel implementations no easier than before; deployment is easier MPI is wonderful; it will be ignored in real world unless simplified CCI notes MPI is HPCC Ghetto CCI is high performance distributed message passing ghetto? CCR from Microsoft – only ~7 primitives – is one possible commodity multicore driver It is roughly active messages Will run MPI style codes fine on multicore 10

Parallel Overhead = (PT(P)/T(1)-1) On P processors = (1/efficiency)-1 CCR Threads per Process Nodes MPI Processes per Node way 16-way 8-way 4-way 2-way Deterministic Annealing Clustering Scaled Speedup Tests on 4 8-core Systems 1,600,000 points per C# thread 1, 2, 4. 8, 16, 32-way parallelism

Overhead PT(P)/T(1)-1 of the messaging runtime for the different data sizes All perform well for large enough datasets Number of Data Points MPI MR Java MR Java MPI

HADOOP MPI In memory MapReduce Factor of 10 3 Factor of 30 Number of Data Points

N=3000 sequences each length ~1000 features Only use pairwise distances will repeat with 0.1 to 0.5 million sequences with a larger machine C# with CCR and MPI 14