Hao Wu, Shangping Ren, Gabriele Garzoglio, Steven Timm, Gerard Bernabeu, Hyun Woo Kim, Keith Chadwick, Seo-Young Noh A Reference Model for Virtual Machine.

Slides:



Advertisements
Similar presentations
Sponsors and Acknowledgments This work is supported in part by the National Science Foundation under Grants No. OCI , IIP and CNS
Advertisements

Fermilab, the Grid & Cloud Computing Department and the KISTI Collaboration GSDC - KISTI Workshop Jangsan-ri, Nov 4, 2011 Gabriele Garzoglio Grid & Cloud.
SLA-Oriented Resource Provisioning for Cloud Computing
Adopting Provenance-based Access Control in OpenStack Cloud IaaS October, 2014 NSS Presentation Institute for Cyber Security University of Texas at San.
Locality-Aware Dynamic VM Reconfiguration on MapReduce Clouds Jongse Park, Daewoo Lee, Bokyeong Kim, Jaehyuk Huh, Seungryoul Maeng.
GPU Virtualization Support in Cloud System Ching-Chi Lin Institute of Information Science, Academia Sinica Department of Computer Science and Information.
The Who, What, Why and How of High Performance Computing Applications in the Cloud Soheila Abrishami 1.
Energy-efficient Virtual Machine Provision Algorithms for Cloud System Ching-Chi Lin Institute of Information Science, Academia Sinica Department of Computer.
1 Distributed Systems Meet Economics: Pricing in Cloud Computing Hadi Salimi Distributed Systems Lab, School of Computer Engineering, Iran University of.
An Approach to Secure Cloud Computing Architectures By Y. Serge Joseph FAU security Group February 24th, 2011.
XENMON: QOS MONITORING AND PERFORMANCE PROFILING TOOL Diwaker Gupta, Rob Gardner, Ludmila Cherkasova 1.
Automatic Resource Scaling for Web Applications in the Cloud Ching-Chi Lin Institute of Information Science, Academia Sinica Department of Computer Science.
Evaluating GPU Passthrough in Xen for High Performance Cloud Computing Andrew J. Younge 1, John Paul Walters 2, Stephen P. Crago 2, and Geoffrey C. Fox.
Towards High-Availability for IP Telephony using Virtual Machines Devdutt Patnaik, Ashish Bijlani and Vishal K Singh.
VMFlock: VM Co-Migration Appliance for the Cloud Samer Al-Kiswany With: Dinesh Subhraveti Prasenjit Sarkar Matei Ripeanu.
DESIGN CONSIDERATIONS OF A GEOGRAPHICALLY DISTRIBUTED IAAS CLOUD ARCHITECTURE CS 595 LECTURE 10 3/20/2015.
Cloud Computing (101).
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 4.
CloudCmp: Shopping for a Cloud Made Easy Ang Li Xiaowei Yang Duke University Srikanth Kandula Ming Zhang Microsoft Research 6/22/2010HotCloud 2010, Boston1.
Plan Introduction What is Cloud Computing?
Measuring zSeries System Performance Dr. Chu J. Jong School of Information Technology Illinois State University 06/11/2012 Sponsored in part by Deer &
Cloud Computing Why is it called the cloud?.
Building Sustainable MIS Infrastuctures
Abstract Cloud data center management is a key problem due to the numerous and heterogeneous strategies that can be applied, ranging from the VM placement.
Introduction to Cloud Computing
Utility Computing Casey Rathbone 1http://cyberaide.org.edu.
Customized cloud platform for computing on your terms !
Cloud Computing. What is Cloud Computing? Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable.
A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.
Introduction to Cloud Computing
M.A.Doman Short video intro Model for enabling the delivery of computing as a SERVICE.
Storage Management in Virtualized Cloud Environments Sankaran Sivathanu, Ling Liu, Mei Yiduo and Xing Pu Student Workshop on Frontiers of Cloud Computing,
Improving Network I/O Virtualization for Cloud Computing.
The Only Constant is Change: Incorporating Time-Varying Bandwidth Reservations in Data Centers Di Xie, Ning Ding, Y. Charlie Hu, Ramana Kompella 1.
IISWC 2007 Panel Benchmarking in the Web 2.0 Era Prashant Shenoy UMass Amherst.
CloudCast: Cloud Computing for Short-term Mobile Weather Forecasts Dilip Kumar Krishnappa, David Irwin, Eric Lyons and Michael Zink IPCCC 2012.
Dynamic Resource Monitoring and Allocation in a virtualized environment.
Sandor Acs 05/07/
The Grid & Cloud Computing Department at Fermilab and the KISTI Collaboration Meeting with KISTI Nov 1, 2011 Gabriele Garzoglio Grid & Cloud Computing.
The New Zealand Institute for Plant & Food Research Limited Use of Cloud computing in impact assessment of climate change Kwang Soo Kim and Doug MacKenzie.
Biomedical Big Data Training Collaborative biobigdata.ucsd.edu BBDTC UPDATES Biomedical Big Data Training Collaborative biobigdata.ucsd.edu.
FermiCloud: Enabling Scientific Workflows with Federation and Interoperability Steven C. Timm FermiCloud Project Lead Grid & Cloud Computing Department.
Feifei Chen Swinburne University of Technology Melbourne, Australia
BlueArc IOZone Root Benchmark How well do VM clients perform vs. Bare Metal clients? Bare Metal Reads are (~10%) faster than VM Reads. Bare Metal Writes.
Accounting for Load Variation in Energy-Efficient Data Centers
Web Technologies Lecture 13 Introduction to cloud computing.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
OpenNebula: Experience at SZTAKI Peter Kacsuk, Sandor Acs, Mark Gergely, Jozsef Kovacs MTA SZTAKI EGI CF Helsinki.
KAASHIV INFOTECH – A SOFTWARE CUM RESEARCH COMPANY IN ELECTRONICS, ELECTRICAL, CIVIL AND MECHANICAL AREAS
Auxiliary services Web page Secrets repository RSV Nagios Monitoring Ganglia NIS server Syslog Forward FermiCloud: A private cloud to support Fermilab.
Fermilab / FermiGrid / FermiCloud Security Update Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359 Keith Chadwick Grid.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
Authentication, Authorization, and Contextualization in FermiCloud S. Timm, D. Yocum, F. Lowe, K. Chadwick, G. Garzoglio, D. Strain, D. Dykstra, T. Hesselroth.
© 2012 Eucalyptus Systems, Inc. Cloud Computing Introduction Eucalyptus Education Services 2.
On-demand Services for the Scientific Program at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab.
GlobusWorld 2012: Experience with EXPERIENCE WITH GLOBUS ONLINE AT FERMILAB Gabriele Garzoglio Computing Sector Fermi National Accelerator.
Fabric for Frontier Experiments at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab ISGC – Thu,
Enabling On-Demand Scientific Workflows on a Federated Cloud Steven C. Timm, Gabriele Garzoglio, Seo-Young Noh, Haeng-Jin Jang KISTI project evaluation.
Parag Mhashilkar (Fermi National Accelerator Laboratory)
FermiCloud Review Response to Questions Keith Chadwick Steve Timm Gabriele Garzoglio Work supported by the U.S. Department of Energy under contract No.
Daniele Lezzi Execution of scientific workflows on federated multi-cloud infrastructures IBERGrid Madrid, 20 September 2013.
Unit 3 Virtualization.
Chapter 6: Securing the Cloud
LIGHTWEIGHT CLOUD COMPUTING FOR FAULT-TOLERANT DATA STORAGE MANAGEMENT
Distributed Network Traffic Feature Extraction for a Real-time IDS
Provisioning Performance of name server Software
Public vs Private Cloud Usage Costs:
Sky Computing on FutureGrid and Grid’5000
A workload-aware energy model for VM migration
Sky Computing on FutureGrid and Grid’5000
Presentation transcript:

Hao Wu, Shangping Ren, Gabriele Garzoglio, Steven Timm, Gerard Bernabeu, Hyun Woo Kim, Keith Chadwick, Seo-Young Noh A Reference Model for Virtual Machine Launching Overhead Presenter: Hao Wu 2014/5/28 Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359 And by joint CRADA FRA / KISTI C13013 between KISTI and Fermilab And supported in part by NSF under grant number CAREER and CNS

Outline 2  Introduction FermiCloud Problem Statement  Observation Baseline CPU Utilization Impact IO Utilization Impact Network Utilization Impact Image Repository Impact  VM Launching Overhead Model  Evaluation  Conclusion

Introduction: Fermilab Fermi National Accelerator Laboratory:  Leading United States particle physics laboratory  Data center for: Energy Frontier (CDF, D0, CMS) Cosmic Frontier (DES, LSST, Darkside, etc.) Intensity Frontier (NOvA, LBNE, Mu2e, MINOS, etc.) 3

Introduction: FermiCloud FermiCloud Project was established in 2009 with the goal of developing and establishing Scientific Cloud capabilities for the Fermilab Scientific Program. FermiCloud Infrastructure as a Service (IaaS) running since FermiCloud Project now focusing on On-demand Services for Scientific Users (PaaS, SaaS). 4

Introduction: Cloud Bursting Cloud bursting tool: GlideinWMS, vcluster Automatically allocate resources on collaborative private cloud (gCloud, etc. ), community clouds (Rackspace, etc. ), and public clouds (Amazon EC2, etc.) 3W challenges: 1.What to burst? 2.When to burst? 3.Where to burst? 5

Introduction: Problem Statement VM launching overhead in literature: Constant Negligible FermiCloud daily operation: Launch time has large variation Consume CPU utilization and IO utilization 6

7 Example Scenarios Scenario one:  Only one host available in private cloud  Several VMs are being launched on the host  A new VM is launched Scenario two:  A task needs 5 more minutes to complete  A new VM needed for a new task  VM launch overhead is 6 minutes Without a reference model, VM launching process can lead resource and cost waste!!

System Architecture System Configuration: Front-end: 16-core Intel(R)Xeon(R) CPU 2.67GHz, 48GB memory. 15 VM hosts: 8-core Intel(R) Xeon(R) CPU 2.66GHz and 16GB memory Cloud Platform: OpenNebula 8

9 OpenNebula VM Launching Process

10 Methodology System information collection : Noninvasive programs, i.e., Iostat, sar VM launch time: Retrieve information from VM system log Start time of SSHD service System CPU Utilization control: cgroup

11 Terminology System CPU Utilization Total CPU utilizations consumed by all the processes on one host System IO Utilization Total IO utilizations consumed by all the processes on one host VM CPU Utilization: CPU utilization consumed by a virtual machine on a single core Prologue (Prolog) The process of copying an image from image repository to the host machine

12 Baseline VM Launching Overhead VM configuration: one virtual core, 2GB memory, 16GB raw image Prolog (scp) Boot

13 System CPU Utilization Impact VMs are launched under different system CPU utilization

14 System IO Utilization Impact VMs are launched under different system IO utilization

15 Simultaneous Launches 2 & 3 VMs are launched simultaneously

16 Network Traffic Impact VMs are launched under different system network utilization

17 Image Repository Impact VMs are launched on different hosts simultaneously

18 Summary of observation VM launching overhead: Prolog (image copying/transferring) Booting overhead Prolog: Significant variations Booting overhead: Steady, i.e. has less variations

19 Notations

20 Modeling Overhead on Prolog Process

21 Modeling Overhead on Booting Process

22 VM Launching Overhead

23 Modeling System Utilization Consumption

Outline IIntroduction FermiCloud Problem Statement OObservation Baseline CPU Utilization Impact IO Utilization Impact Network Utilization Impact Image Repository Impact VVM Launching Overhead Model EEvaluation CConclusion 24

25 Modeling Overhead

26 Evaluation: Baseline

27 Evaluation

28 Evaluation: 3 Simultaneous Launches

29 Evaluation: Random Launches

30 Evaluation: Overall Performance Utilization DifferenceTime Difference Baseline0.03%2.3% 2 Sim. Launches4.46%5.1% 3 Sim. Launches6.28%1.7% Random. Launches4.91%6.9% Overall3.92%4.0%

Outline IIntroduction FermiCloud Problem Statement OObservation Baseline CPU Utilization Impact IO Utilization Impact Network Utilization Impact Image Repository Impact VVM Launching Overhead Model EEvaluation CConclusion 31

32 Conclusion Problem Addressed: VM launching overhead has large variation Without a VM launching overhead reference may lead to resource and cost waste Contribution: Proposed a VM launching overhead reference model based on real FermiCloud operational data CPU overhead IO overhead Timing overhead Evaluation indicates the proposed model can fairly predict VM launching overhead

33 Acknowledgement

34 Thanks! Questions?