Storage Management in Virtualized Cloud Environments Sankaran Sivathanu, Ling Liu, Mei Yiduo and Xing Pu Student Workshop on Frontiers of Cloud Computing,

Slides:



Advertisements
Similar presentations
Virtual Memory (II) CSCI 444/544 Operating Systems Fall 2008.
Advertisements

Capacity Planning in a Virtual Environment
SLA-Oriented Resource Provisioning for Cloud Computing
The Who, What, Why and How of High Performance Computing Applications in the Cloud Soheila Abrishami 1.
© 2005 DataCore Software Corp SANmelody™ 2.0 IP/SAN FC/SAN Application Support Services Thin Provisioning Automation & Disk Resource Consolidation DataCore,
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
Introduction CSCI 444/544 Operating Systems Fall 2008.
VSphere vs. Hyper-V Metron Performance Showdown. Objectives Architecture Available metrics Challenges in virtual environments Test environment and methods.
1 Distributed Systems Meet Economics: Pricing in Cloud Computing Hadi Salimi Distributed Systems Lab, School of Computer Engineering, Iran University of.
XENMON: QOS MONITORING AND PERFORMANCE PROFILING TOOL Diwaker Gupta, Rob Gardner, Ludmila Cherkasova 1.
What will my performance be? Resource Advisor for DB admins Dushyanth Narayanan, Paul Barham Microsoft Research, Cambridge Eno Thereska, Anastassia Ailamaki.
NoHype: Virtualized Cloud Infrastructure without the Virtualization Eric Keller, Jakub Szefer, Jennifer Rexford, Ruby Lee ISCA 2010 Princeton University.
SLA-aware Virtual Resource Management for Cloud Infrastructures
Towards High-Availability for IP Telephony using Virtual Machines Devdutt Patnaik, Ashish Bijlani and Vishal K Singh.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
Microsoft Virtual Server 2005 Product Overview Mikael Nyström – TrueSec AB MVP Windows Server – Setup/Deployment Mikael Nyström – TrueSec AB MVP Windows.
ENFORCING PERFORMANCE ISOLATION ACROSS VIRTUAL MACHINES IN XEN Diwaker Gupta, Ludmila Cherkasova, Rob Gardner, Amin Vahdat Middleware '06 Proceedings of.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Virtualization in Data Centers Prashant Shenoy
DESIGN CONSIDERATIONS OF A GEOGRAPHICALLY DISTRIBUTED IAAS CLOUD ARCHITECTURE CS 595 LECTURE 10 3/20/2015.
STORAGE Virtualization
Distributed Systems Meet Economics: Pricing In The Cloud Authors: Hongyi Wang, Qingfeng Jing, Rishan Chen, Bingsheng He, Zhengping He, Lidong Zhou Presenter:
New Challenges in Cloud Datacenter Monitoring and Management
Virtualization Performance H. Reza Taheri Senior Staff Eng. VMware.
VMware vSphere 4 Introduction. Agenda VMware vSphere Virtualization Technology vMotion Storage vMotion Snapshot High Availability DRS Resource Pools Monitoring.
Measuring zSeries System Performance Dr. Chu J. Jong School of Information Technology Illinois State University 06/11/2012 Sponsored in part by Deer &
Generating Adaptation Policies for Multi-Tier Applications in Consolidated Server Environments College of Computing Georgia Institute of Technology Gueyoung.
Dual Stack Virtualization: Consolidating HPC and commodity workloads in the cloud Brian Kocoloski, Jiannan Ouyang, Jack Lange University of Pittsburgh.
Dynamic and Decentralized Approaches for Optimal Allocation of Multiple Resources in Virtualized Data Centers Wei Chen, Samuel Hargrove, Heh Miao, Liang.
Key Perf considerations & bottlenecks Windows Azure VM characteristics Monitoring TroubleshootingBest practices.
Introduction to VMware Virtualization
Virtual Machine Course Rofideh Hadighi University of Science and Technology of Mazandaran, 31 Dec 2009.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
Cloud Computing Energy efficient cloud computing Keke Chen.
Adaptive software in cloud computing Marin Litoiu York University Canada.
Improving Network I/O Virtualization for Cloud Computing.
Improving Disk Latency and Throughput with VMware Presented by Raxco Software, Inc. March 11, 2011.
An Autonomic Framework in Cloud Environment Jiedan Zhu Advisor: Prof. Gagan Agrawal.
Challenges towards Elastic Power Management in Internet Data Center.
MClock: Handling Throughput Variability for Hypervisor IO Scheduling in USENIX conference on Operating Systems Design and Implementation (OSDI ) 2010.
Eneryg Efficiency for MapReduce Workloads: An Indepth Study Boliang Feng Renmin University of China Dec 19.
DBI313. MetricOLTPDWLog Read/Write mixMostly reads, smaller # of rows at a time Scan intensive, large portions of data at a time, bulk loading Mostly.
Quantifying and Improving I/O Predictability in Virtualized Systems Cheng Li, Inigo Goiri, Abhishek Bhattacharjee, Ricardo Bianchini, Thu D. Nguyen 1.
Hakim Weatherspoon Robbert van Renesse SUPERCLOUD: GOING BEYOND FEDERATED CLOUDS 1.
Xiao Ling 1, Shadi Ibrahim 2, Hai Jin 1, Song Wu 1, Songqiao Tao 1 1 Cluster and Grid Computing Lab Services Computing Technology and System Lab School.
Providing Differentiated Levels of Service in Web Content Hosting Jussara Almeida, etc... First Workshop on Internet Server Performance, 1998 Computer.
VGreen: A System for Energy Efficient Manager in Virtualized Environments G. Dhiman, G Marchetti, T Rosing ISLPED 2009.
Consolidation and Optimization Best Practices: SQL Server 2008 and Hyper-V Dandy Weyn | Microsoft Corp. Antwerp, March
VMware vSphere Configuration and Management v6
Virtualization and Databases Ashraf Aboulnaga University of Waterloo.
June 30 - July 2, 2009AIMS 2009 Towards Energy Efficient Change Management in A Cloud Computing Environment: A Pro-Active Approach H. AbdelSalamK. Maly.
Cloud Computing Lecture 5-6 Muhammad Ahmad Jan.
Web Technologies Lecture 13 Introduction to cloud computing.
PROOF Benchmark on Different Hardware Configurations 1 11/29/2007 Neng Xu, University of Wisconsin-Madison Mengmeng Chen, Annabelle Leung, Bruce Mellado,
Capacity Planning in a Virtual Environment Chris Chesley, Sr. Systems Engineer
Virtual Machine in HPC PAK MARKTHUB (13M54040) 1 VIRTUAL MACHINE IN HPC.
REMINDER Check in on the COLLABORATE mobile app Best Practices for Oracle on VMware - Deep Dive Darryl Smith Chief Database Architect Distinguished Engineer.
Architecture for Resource Allocation Services Supporting Interactive Remote Desktop Sessions in Utility Grids Vanish Talwar, HP Labs Bikash Agarwalla,
© 2012 Eucalyptus Systems, Inc. Cloud Computing Introduction Eucalyptus Education Services 2.
1 Automated Power Management Through Virtualization Anne Holler, VMware Anil Kapur, VMware.
Organizations Are Embracing New Opportunities
Introduction to VMware Virtualization
Amazon Instance Purchasing Options
PA an Coordinated Memory Caching for Parallel Jobs
Comparison of the Three CPU Schedulers in Xen
Zhen Xiao, Qi Chen, and Haipeng Luo May 2013
IIS Progress Report 2016/01/18.
Exploring Multi-Core on
Xing Pu21 Ling Liu1 Yiduo Mei31 Sankaran Sivathanu1 Younggyun Koh1
Presentation transcript:

Storage Management in Virtualized Cloud Environments Sankaran Sivathanu, Ling Liu, Mei Yiduo and Xing Pu Student Workshop on Frontiers of Cloud Computing, IBM 2010

2 Talk Outline Introduction Measurement results & Observations –Data Placement & Provisioning –Workload Interference –Impacts of Virtualization Summary

3 Cloud & Virtualization Cloud Environment – Goals –Flexibility in resource configuration –Maximum resource utilization –Pay-per-use Model Virtualization – Benefits –Resource consolidation –Re-structuring flexibility –Separate protection domains Virtualization suits as one of the basic foundations of Cloud infrastructures

4 Fundamental Issues Could Service Providers (CSPs) vs. Customers –Customers purchase computing resources –CSPs provide virtual resources (VMs) –Customers perceive their resources as physical machines! Multiple VMs reside in single physical host –Resource Interference –End-user performance depends on other users End-user unaware of where their data physically exists

5 Goals of our Measurement For cloud service providers –How to place data such that end-user performance is maximized ? –How to co-locate workloads for least interference ? For End-Users –How to purchase resources in tune with requirement ? –How to tune applications for maximum performance ? General insights on storage I/O in virtualized environments

6 Benchmarks Used Postmark –Mail Server Workload –Create/Delete, Read/Append files –Parameters File Size # of files Read/Write ratio Synthetic Workload –Sequential vs. random accesses –Zipf Distribution

7 Data Provisioning & Placement

8 Workload Data footprint ~150MB 4GB Partition 40GB Partition Throughput : 2.1 MB/s Throughput : 1.4 MB/s Performance Difference : 33% Disk Provisioning Consider a 100GB Disk Case - ICase - II

9 Where to place VM disk ? Postmark benchmark –Read operation Cases : –Read from physical partitions in different zones Based on LBNs LBNs start from inner zone and proceeds towards outer zones. –Read from disk file (.vmdk)

10 Where to place multiple VM disks ? Postmark benchmark –2 instances (1 for each VM) Random reads Compare physical partitions placed in different zones –O -> Outer –I -> Inner –M -> Mid

11 Observations Customers should purchase storage based on workload requirement, not price Thin provisioning may be practiced Throughput intensive VMs can be placed in outer disk zones Multiple VMs that may be accessed simultaneously should be co-located on disk –CSPs can monitor access patterns and move virtual disks accordingly

12 Workload Interference

13 CPU-Disk Interference VM - 1VM - 2 CPU DISK Throughput : 23.4 MB/s CPU Throughput : 27.6 MB/s Performance Difference : 15.3% Physical Host

14 CPU-Disk Interference CPU allocation ratios has no effect on disk throughput across VMs Disk intensive job performs better along with a CPU intensive job

15 Reason ? Dynamic Frequency Scaling CPU-Disk Interference

16 CPU-Disk Interference CPU DFS is enabled in Linux by default Three ‘governors’ to control the DFS policy On-demand (default) Performance Power-save When 1 core is idle, entire CPU is down-scaled because overall CPU utilization falls

17 Disk-Disk Interference 1 instance of Postmark in each VMs 65.3% more time taken when compared to running Postmark in a single VM Overhead mainly attributed to disk seeks : No more sequential accesses CPU V.Disk-1 V.Disk-2 Physical Disk VM-1 VM-2 Physical Host

18 CPU V.Disk-1 V.Disk-2 Disk - 1Disk - 2 VM-1 VM-2 Disk-Disk Interference VMs using separate physical disks 17.52% more time taken when compared to running Postmark in a single VM Overhead attributed to contention in Dom-0’s queue structures Physical Host

19 Disk-Disk Interference Postmark Benchmark (Reads) Cases : –Running in a single VM –1 instance in each of two VMs 2 VMs reading from virtual disks in same physical disk 2 VMs reading from virtual disks in different physical disks

20 Disk-Disk Interference IO scheduling policy in Dom-0 has less effect ‘Ideal’ case is time taken when running Postmark in single VM Other cases are running 1 instance of Postmark in each of 2 VMs (separate physical disks)

21 Disk-Disk Interference Interference with respect to workload type Synthetic read workload VMs use separate physical disks Cases : –Mix of sequential versus random reads Sequential requests from both VMs flood Dom-0 queue - contention

22 Observations CPU-intensive and disk-intensive workloads can be co- located for optimal performance and power Virtual disks that may be accessed simultaneously must be placed in separate physical disks I/O scheduling in Dom-0 has less effect on disk workload interference Two sequential workloads, when co-located suffer in performance due to queue contention With separate disks, workload contention is generally minimal, other than the case of two sequential workloads

23 Impacts of Virtualization

24 Sequentiality Postmark benchmark (reads) No much overhead seen for random disk accesses VM overhead is mitigated by larger disk overhead More felt for sequential disk accesses

25 Block Size Postmark sequential reads Fixed overhead with every requests As block sizes increase, # of requests are reduced, hence overhead is reduced Efficient to read in larger blocks

26 Block size wrt. Locality

27 Observations VM overhead is not felt in random workloads – amortized by disk seeks Extra layers of indirection is the reason for VM overhead – when block size is large, overhead is amortized Block size may be increased only if there is sufficient locality in access

28 Summary Storage purchased must depend on requirement, not price! It is better to place sequentially accessed streams in outer disk zone Co-locate virtual disks that may be accessed simultaneously Co-locate CPU intensive task with disk intensive task for better power and performance Avoid co-locating two sequential workloads on single physical machine – even when it goes to separate physical disks! Read in large blocks only when there is locality in workload

29 Questions Contact :