Cloud Computing: Is It Powerful, And Can It Lead To Bankruptcy? G. Bruce Berriman Infrared Processing.

Slides:



Advertisements
Similar presentations
University of Notre Dame
Advertisements

Cloud Computing Imranul Hoque. Today’s Cloud Computing.
Approaches to Investigating Technology Solutions In Data Intensive Astronomy G. Bruce Berriman NASA Exoplanet Science Institute, Infrared.
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Clouds C. Vuerli Contributed by Zsolt Nemeth. As it started.
Ken Birman. Massive data centers We’ve discussed the emergence of massive data centers associated with web applications and cloud computing Generally.
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
An Astronomical Image Mosaic Service for the National Virtual Observatory / ESTO.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 4.
Cloud computing Tahani aljehani.
An Introduction to Cloud Computing. The challenge Add new services for your users quickly and cost effectively.
Plan Introduction What is Cloud Computing?
CREATING A MULTI-WAVELENGTH GALACTIC PLANE ATLAS WITH AMAZON WEB SERVICES G. Bruce Berriman, John Good IPAC, California Institute of Technolog y Ewa Deelman,
Scientific Workflows in the Cloud Gideon Juve USC Information Sciences Institute
Cloud Computing Why is it called the cloud?.
Clouds on IT horizon Faculty of Maritime Studies University of Rijeka Sanja Mohorovičić INFuture 2009, Zagreb, 5 November 2009.
Utility Computing Casey Rathbone 1http://cyberaide.org.edu.
Ch 4. The Evolution of Analytic Scalability
Scientific Workflows in the Cloud Gideon Juve USC Information Sciences Institute
A Very Brief Introduction To Cloud Computing Jens Vöckler, Gideon Juve, Ewa Deelman, G. Bruce Berriman.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
Adam Leidigh Brandon Pyle Bernardo Ruiz Daniel Nakamura Arianna Campos.
Cloud Computing Saneel Bidaye uni-slb2181. What is Cloud Computing? Cloud Computing refers to both the applications delivered as services over the Internet.
PhD course - Milan, March /09/ Some additional words about cloud computing Lionel Brunie National Institute of Applied Science (INSA) LIRIS.
Cloud computing is the use of computing resources (hardware and software) that are delivered as a service over the Internet. Cloud is the metaphor for.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 7 2/23/2015.
Cloud Computing. What is Cloud Computing? Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Software Architecture
DISTRIBUTED COMPUTING
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
Introduction to Cloud Computing
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Running Scientific Workflow Applications on the Amazon EC2 Cloud Bruce Berriman NASA Exoplanet Science Institute, IPAC Gideon Juve, Ewa Deelman, Karan.
Cloud Computing Dave Elliman 11/10/2015G53ELC 1. Source: NY Times (6/14/2006) The datacenter is the computer!
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Experiences Using Cloud Computing for A Scientific Workflow Application Jens Vöckler, Gideon Juve, Ewa Deelman, Mats Rynge, G. Bruce Berriman Funded by.
Benchmarking MapReduce-Style Parallel Computing Randal E. Bryant Carnegie Mellon University.
Ten Years of Software Sustainability at The Infrared Processing and Analysis Center G. Bruce Berriman and John Good NASA Exoplanet Science Institute, Infrared.
The New Zealand Institute for Plant & Food Research Limited Use of Cloud computing in impact assessment of climate change Kwang Soo Kim and Doug MacKenzie.
Chapter 8 – Cloud Computing
Carnegie Mellon University © Robert T. Monroe Management Information Systems Cloud Computing I Cloud Models and Technologies Management.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-1.
Web Technologies Lecture 13 Introduction to cloud computing.
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-2.
1 TCS Confidential. 2 Objective : In this session we will be able to learn:  What is Cloud Computing?  Characteristics  Cloud Flavors  Cloud Deployment.
RANDY MODOWSKI COSC Cloud Computing. Road Map What is Cloud Computing? History of “The Cloud” Cloud Milestones How Cloud Computing is being used.
KAASHIV INFOTECH – A SOFTWARE CUM RESEARCH COMPANY IN ELECTRONICS, ELECTRICAL, CIVIL AND MECHANICAL AREAS
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
© 2012 Eucalyptus Systems, Inc. Cloud Computing Introduction Eucalyptus Education Services 2.
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
Agenda  What is Cloud Computing?  Milestone of Cloud Computing  Common Attributes of Cloud Computing  Cloud Service Layers  Cloud Implementation.
HPC In The Cloud Case Study: Proteomics Workflow
Unit 3 Virtualization.
Accessing the VI-SEEM infrastructure
Organizations Are Embracing New Opportunities
Cloudy Skies: Astronomy and Utility Computing
Seismic Hazard Analysis Using Distributed Workflows
Grid Computing.
Recap: introduction to e-science
Introduction to Cloud Computing
CNIT131 Internet Basics & Beginning HTML
INFO 344 Web Tools And Development
Ch 4. The Evolution of Analytic Scalability
Distributing META-pipe on ELIXIR compute resources
Presentation transcript:

Cloud Computing: Is It Powerful, And Can It Lead To Bankruptcy? G. Bruce Berriman Infrared Processing and Analysis Center, Caltech Space Telescope Science Institute, February

Collaborators Ewa Deelman Gideon Juve Mats Rynge Jens-S Völcker Information Sciences Institute, USC 2

Cloud Computing Is Not A New Idea! 3  Rewind to the early1960’s ….  John McCarthy … “computation delivered as a public utility in…. the same way as water and power.”  J. C. R. Licklider … “the intergalactic computer network” “It seems to me to be interesting and important to develop a capability for integrated network operation … we would have …perhaps six or eight small computers, and a great assortment of disc files and magnetic tape units … all churning away”

The Idea Was Dormant For 35 Years  … until the internet started to offer significant bandwidth.  Salesforce.com, Amazon Web Services started to offer applications over the internet ( ).  Amazon Elastic Compute cloud (2006) offered first widely accessible on-demand computing.  By 2009, browser based apps such as Google Apps had hit their stride. 4

Cloud Service Layers 5

Developing A New Business Model For Astronomical Computing  Astronomy is already a data intensive science  Over 1 PB served electronically through data centers and archives.  By 2020, as much as PB on- line.  Astro2010 recognized the need for demand high performance computing on massive, distributed data sets. 6

How Useful Is Cloud Computing For Scientific Workflow Applications? The study was designed to answer the question: How useful is cloud computing for scientific workflow applications ? Workflow applications are loosely coupled applications in which the output files from one component become the input to the next. There were three goals: 1. Conduct an experimental study of the performance of three workflows with different I/O, memory and CPU requirements on a commercial cloud. 2. Compare the performance of cloud resources with the performance of a typical High Performance Cluster (HPC). The cloud uses commodity hardware and virtualization and HPCs use parallel file systems and fast networks. 3. Provide a n analysis of the various costs associated with running workflows on a commercial cloud. We chose Amazon EC2 as the cloud provider and the NCSA Abe cluster as a high-performance cluster. 1. What is performance/cost of different cloud resource configurations for workflow applications? 2. What is performance of cloud and workflow applications? 3. What is virtualization overhead on cloud?  Loosely-coupled parallel applications  Many domains: astronomy, biology, earth science.  Potentially very large: 10 K tasks common, >1M not uncommon  Potentially data-intensive: 10 GB common, >1TB not uncommon  Data communicated via files  Shared storage system, or network transfers 7

Getting Started With Cloud Computing All you need is a credit card and connect to Amazon EC2 front page: Commercial Providers Amazon.com EC2 AT&T Synaptic Hosting GNi Dedicated Hosting IBM Computing on Demand Rackspace Cloud Servers Savvis Open Cloud ServePath GoGrid Skytap Virtual Lab 3Tera Unisys Secure Verizon Computing Zimory Gateway Science Clouds FutureGrid NERSC Magellan NASA Nebula This looks cheap! 8

“Little sins add up …” … and that’s not all. You pay for: -Transferring data into the cloud -Transferring them back out again -Storage while you are processing (or sitting idle) -Storage of the VM and your own software -Special services: virtual private cloud… See Manav Gupta’s blog post Annual Costs! 9

What Amazon EC2 Does  Creates as many independent virtual machines as you wish.  Reserves the storage space you need.  Gives you a refund if their equipment fails.  Bills you. 10

What You Have To Do  Configure the virtual machines and create your environment  Load all your software and input data  Manage and maintain 11  Working at scale:  Adapt applications to new computing models in cloud (e.g. MapReduce) or  Adapt the cloud to recreate the environment in which the app has run - a virtual cluster.

Computational Models: MapReduce 12  Wiley et al (2011) “Astronomy in the Cloud: Using MapReduce for Image Co-Addition” PASP, 123, 366.  SDSS Image co-addition of 20 TB data  100,000 files processed in 3 minutes on 400 cores  Considerable effort to adapt co-addition to Hadoop Map : partitions input into smaller sub-problems, and distributes them to worker nodes. Reduce: collects answers to sub-problems and combines them to form the output Map : partitions input into smaller sub-problems, and distributes them to worker nodes. Reduce: collects answers to sub-problems and combines them to form the output

Pegasus Workflow Management System  Converts abstract workflow to a concrete workflow  No special requirements on infrastructure 13 Condor Schedd DAGMan Pegasus Maps tasks and data to executable resources; Performance optimizer Cyberinfrastructure: Local machine, cluster, Condor pool, Grid, Cloud,… Task manager; schedules and dispatches tasks (and data) to resources Workflow engine - Tracks dependencies, releases tasks, retries tasks

Comparing Clusters With Clouds NCSA Abe - high- performance cluster. Amazon EC2 14 Set up computationally equivalent configurations on Amazon and Abe Set up computationally equivalent configurations on Amazon and Abe Corral – resource provisioning tool for grids. Nimbus Context Broker – toolkit for configuring virtual clusters.

The Applications Montage ( creates science-grade image mosaics from multiple input images. Broadband simulates and compares seismograms from earthquake simulation codes. Epigenome maps short DNA segments collected using high-throughput gene sequencing machines to a reference genome. Montage ( creates science-grade image mosaics from multiple input images. Broadband calculates seismograms from simulated earthquakes. Epigenome maps short DNA segments collected with gene sequencing machines to a reference genome. Montage ( creates science-grade image mosaics from multiple input images. Broadband calculates seismograms from simulated earthquakes. Epigenome maps short DNA segments collected with gene sequencing machines to a reference genome. Montage Workflow ReprojectionBackground Rectification Co-addition Output Input Montage Workflow ReprojectionBackground Rectification Co-addition Output Input Montage Workflow ReprojectionBackground Rectification Co-addition Output Input Montage Workflow ReprojectionBackground Rectification Co-addition Output Input 15

Characteristics of Workflows Resource Usage of the Three Workflow Applications Workflow Specifications for this Study 16

Computing Resources Processors and OS  Amazon offers wide selection of processors.  Ran Linux Red Hat Enterprise with VMWare  c1.xlarge and abe.local are equivalent – estimate overhead due to virtualization  abe.lustre and abe.local differ only in file system Networks and File Systems  HPC systems use high-performance network and parallel file systems  Amazon EC2 uses commodity hardware  Ran all processes on single, multi- core nodes. Used local and parallel file system on Abe. 17

Performance Results   Virtualization Overhead <10%  Large differences in performance between the resources and between the applications  The parallel file system on abe.lustre offers a big performance advantage of x3 for Montage 18

How Much Did It Cost ? InstanceCost $/hr m1.small0.10 m1.large0.40 m1.xlarge0.80 c1.medium0.20 c1.xlarge0.80 Montage:  Clear trade-off between performance and cost.  Most powerful processor c1.xlarge offers 3x the performance of m1.small – but at 4x the cost.  Most cost-effective processor for Montage is c1.medium – 20% performance loss over m1.small, but 4x lower cost. 19

Data Transfer Costs OperationCost $/GB Transfer In0.10 Transfer Out0.17 ApplicationInput (GB)Output (GB)Logs (MB) Montage Broadband Epigenome ApplicationInputOutputLogsTotal Montage$0.42$1.32<$0.01$1.75 Broadband$0.40$0.03<$0.01$0.43 Epigenome$0.18$0.05<0.01$0.23 Transfer Rates  Amazon charges different rates for transferring data into the cloud and back out again.  Transfer-out costs are the higher of the two. Transfer Costs  For Montage, the cost to transfer data out of the cloud is higher than monthly storage and processing costs.  For Broadband and Epigenome, processing incurs the biggest costs. 20

Storage Costs ItemCharges $ Storage of VM’s in local Disk (S3) 0.15/GB-Month Storage of data in EBS disk0.10/GB-Month Storage Rates Data Storage Charges  Amazon charges for storing Virtual Machines (VM) and user’s applications in local disk  It also charges for storing data in persistent network-attached Elastic Block Storage (EBS). Storage Volumes Storage Costs Montage Storage Costs Exceed Most Cost-Effective Processor Costs 21

The bottom line for Montage ItemBest ValueBest Performance c1.mediumc1.xlarge Transfer Data In$ 0.42 Processing$ 0.55$ 2.45 Storage/month$ 1.07 Transfer Out$ 1.32 Totals$ 3.36$ x the processor cost for 20% better performance 22

Cost-Effective Mosaic Service Local Option Amazon EBS Option Amazon S3 Options Amazon cost is 2X local! -2MASS image data set - 1,000 x 4 square degree mosaics/month 23

Just To Keep It Interesting … Running the Montage Workflow With Different File Storage Systems Cost and performance vary widely with different types of file storage depending on how storage architecture handles lots of small files Cf. Epigenome 24

When Should I Use The Cloud?  The answer is….it depends on your application and use case.  Recommended best practice: Perform a cost-benefit analysis to identify the most cost-effective processing and data storage strategy. Tools to support this would be beneficial.  Amazon offers the best value  For compute- and memory-bound applications.  For one-time bulk-processing tasks, providing excess capacity under load, and running test-beds.  Parallel file systems and high-speed networks offer the best performance for I/O-bound applications.  Mass storage is very expensive on Amazon EC2 25

Hunting Exoplanets with Kepler  Kepler continuously monitors the brightness of over 175,000 stars.  Search for periodic dips in signals as Earth-like planets transit in front of host star.  Currently 380,000 light curves have been released.  20+ new planets have been discovered, nearly 2,000 candidate planets.  Can we perform a bulk analysis of all the data to search for these periodic signals? 26 Kepler 6-b transit

Digging Out Exoplanets with Periodograms  A periodogram calculates the significance of different frequencies in time-series data to identify periodic signals.  NASA Exoplanet Database Periodogram tool  Fast, portable implementation in C  Easily scalable: each frequency sampled independently of all other frequencies  Calculations are slow: 1 hour for 100, ,000 points typical of Kepler light curves.  How can we process the entire data set? Candidate for the cloud: “high-burst,” processor-bound, easily parallelizable. 27

Kepler Periodogram Atlas  Compute periodogram atlas for public Kepler dataset  Use 128 processor cores in parallel on Amazon EC2 and TeraGrid  ~210K light curves X 3 algorithms Estimated cost Compute is ~10X Transfer 28 Amazon: 16 x c1.xlarge instances = 128 cores Ranger: 8-16 x 16 core nodes = cores RunAlgorithmOptimization 1 (EC1)Lomb-ScargleSinusoids 2 (EC1)Box-Least SquaresBox 3 (TG)PlavchanUnrestricted

Digging Out Exoplanets On Academic Clouds 29  FutureGrid test bed for Cloud Computing  6 centers across the U.S.  Nimbus, Eucalyptus 

Computing Periodograms on Academic Clouds 30 SiteCPURAM (SW)WalltimeCum. Dur.Speed-Up Magellan8 x 2.6 GHz19 (0) GB5.2 h226.6 h43.6 Amazon8 x 2.3 GHz7 (0) GB7.2 h295.8 h41.1 FutureGrid8 x 2.5 GHz29 (½) GB5.7 h248.0 h43.5  33 K periodograms with Plavchan algorithm  Given 48 physical cores  Speed-up ≈ 43 considered pretty good  AWS cost ≈ $31:  7.2 h x 6 x c1.large ≈ $29  1.8 GB in GB out ≈ $2  Results encouraging.

Failure To Provision VM’s The Main Problem 31

Summation  Cloud computing is powerful when applied in the right places: “high-burst” applications, especially when they are processor or memory bound.  Be careful of how costs can spiral esp. when storing or moving data.  Always perform a cost-benefit analysis  When mature, academic clouds may offer a realistic alternative to commercial clouds BUT

Caveat Emptor!  “Cloud Computing as it exists today is not ready for High Performance Computing because  Large overheads to convert to Cloud environments  Virtual instances under perform bare-metal systems and  The cloud is less cost-effective than most large centers” Shane Canon et al. (2011). “ Debunking some Common Misconceptions of Science in the Cloud.” Science Cloud Workshop, San Jose, CA.  Similar Conclusions in Magellan Final Report (December 2011 ) 33

Gartner’s Emerging Technologies Hype Cycle 34

CANFAR: Cloud Technology in Action in Astronomy “The Canadian Advanced Network For Astronomical Research (CANFAR) is an operational system for the delivery, processing, storage, analysis, and distribution of very large astronomical datasets. The goal of CANFAR is to support large Canadian astronomy projects.” 35 Nimbus open-source toolkit. IaaS tools to support scientific computing.

 The Application of Cloud Computing to Scientific Workflows: A Study of Cost and Performance. G. Berriman et al Invited Review Paper Submitted to Special e- Science Edition of Philosophical Transactions of the Royal Society A.  Scientific Workflow Applications on Amazon EC2. G. Juve et al. Cloud Computing Workshop in Conjunction with e-Science 2009 (Oxford, UK).  Data Sharing Options for Scientific Workflows on Amazon EC2, G. Juve et al. Proceedings of Supercomputing 10 (SC10),  The Application of Cloud Computing to Astronomy: A Study of Cost and Performance. G. B. Berriman et al Proceedings of “e-Science in Astronomy” Workshop. Brisbane.  Astronomy in the Cloud: Using MapReduce for Image Co-Addition. K. Wiley et al PASP, 123, 366.  Magellan Final Report, December Summary:  Bruce Berriman’s blog, “Astronomy Computing Today,” at Where Can I Learn More? 36