U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Server Consolidation in Virtualized Data Centers Prashant Shenoy University of Massachusetts.

Slides:



Advertisements
Similar presentations
CS 443 Advanced OS Fabián E. Bustamante, Spring 2005 Memory Resource Management in VMware ESX Server Carl A. Waldspurger VMware, Inc. Appears in SOSDI.
Advertisements

Capacity Planning in a Virtual Environment
Virtualisation From the Bottom Up From storage to application.
1 Storage-Aware Caching: Revisiting Caching for Heterogeneous Systems Brian Forney Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Wisconsin Network Disks University.
Difference Engine: Harnessing Memory Redundancy in Virtual Machines by Diwaker Gupta et al. presented by Jonathan Berkhahn.
VSphere vs. Hyper-V Metron Performance Showdown. Objectives Architecture Available metrics Challenges in virtual environments Test environment and methods.
Exploiting Data Deduplication to Accelerate Live Virtual Machine Migration Xiang Zhang 1,2, Zhigang Huo 1, Jie Ma 1, Dan Meng 1 1. National Research Center.
Green Cloud Computing Hadi Salimi Distributed Systems Lab, School of Computer Engineering, Iran University of Science and Technology,
Memory Buddies: Exploiting Page Sharing for Smart Colocation in Virtualized Data Centers Timothy Wood, Gabriel Tarasuk-Levin, Prashant Shenoy, Peter Desnoyers*,
CloudScale: Elastic Resource Scaling for Multi-Tenant Cloud Systems Zhiming Shen, Sethuraman Subbiah, Xiaohui Gu, John Wilkes.
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science Dynamic Provisioning for Multi-tier Internet Applications Bhuvan Urgaonkar, Prashant.
U NIVERSITY OF M ASSACHUSETTS A MHERST Department of Computer Science CRAMM: Virtual Memory Support for Garbage-Collected Applications Ting Yang, Emery.
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science Quantifying the Benefits of Resource Multiplexing in On-Demand Data Centers Abhishek.
OnCall: Defeating Spikes with Dynamic Application Clusters Keith Coleman and James Norris Stanford University June 3, 2003.
U NIVERSITY OF M ASSACHUSETTS Department of Computer Science Automatic Heap Sizing Ting Yang, Matthew Hertz Emery Berger, Eliot Moss University of Massachusetts.
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science Dynamic Resource Allocation for Shared Data Centers Using Online Measurements.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Virtualization in Data Centers Prashant Shenoy
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts, Amherst Operating Systems CMPSCI 377 Lecture.
Computer Science Storage Systems and Sensor Storage Research Overview.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science From Cloud Computing to Sensor Networks: Distributed Computing Research at LASS.
Bandwidth Allocation in a Self-Managing Multimedia File Server Vijay Sundaram and Prashant Shenoy Department of Computer Science University of Massachusetts.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts, Amherst Operating Systems CMPSCI 377 Lecture.
Virtualization for Cloud Computing
Implementing Failover Clustering with Hyper-V
BUFFALO: Bloom Filter Forwarding Architecture for Large Organizations Minlan Yu Princeton University Joint work with Alex Fabrikant,
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Hash, Don’t Cache: Fast Packet Forwarding for Enterprise Edge Routers Minlan Yu Princeton University Joint work with Jennifer.
5205 – IT Service Delivery and Support
Computer Science Cataclysm: Policing Extreme Overloads in Internet Applications Bhuvan Urgaonkar and Prashant Shenoy University of Massachusetts.
VMware vSphere 4 Introduction. Agenda VMware vSphere Virtualization Technology vMotion Storage vMotion Snapshot High Availability DRS Resource Pools Monitoring.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Black-box and Gray-box Strategies for Virtual Machine Migration Timothy Wood, Prashant.
Black-box and Gray-box Strategies for Virtual Machine Migration Timothy Wood, Prashant Shenoy, Arun Venkataramani, and Mazin Yousif † Univ. of Massachusetts.
Department of Computer Science Engineering SRM University
Virtual Machine Hosting for Networked Clusters: Building the Foundations for “Autonomic” Orchestration Based on paper by Laura Grit, David Irwin, Aydan.
How to Resolve Bottlenecks and Optimize your Virtual Environment Chris Chesley, Sr. Systems Engineer
SAIGONTECH COPPERATIVE EDUCATION NETWORKING Spring 2009 Seminar #1 VIRTUALIZATION EVERYWHERE.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science An Analytical Model for Multi-tier Internet Services and its Applications Bhuvan.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 2.
Improving Network I/O Virtualization for Cloud Computing.
1 Geiger: Monitoring the Buffer Cache in a Virtual Machine Environment Stephen T. Jones Andrea C. Arpaci-Dusseau Remzi H. Arpaci-Dusseau Department of.
Kinshuk Govil, Dan Teodosiu*, Yongqiang Huang, and Mendel Rosenblum
Autonomic SLA-driven Provisioning for Cloud Applications Nicolas Bonvin, Thanasis Papaioannou, Karl Aberer Presented by Ismail Alan.
COMS E Cloud Computing and Data Center Networking Sambit Sahu
© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Profiling and Modeling Resource Usage.
Virtualization Part 2 – VMware. Virtualization 2 CS5204 – Operating Systems VMware: binary translation Hypervisor VMM Base Functionality (e.g. scheduling)
Our work on virtualization Chen Haogang, Wang Xiaolin {hchen, Institute of Network and Information Systems School of Electrical Engineering.
May 30, 2016Department of Computer Sciences, UT Austin1 Using Bloom Filters to Refine Web Search Results Navendu Jain Mike Dahlin University of Texas at.
Server Virtualization
1 Virtual Machine Memory Access Tracing With Hypervisor Exclusive Cache USENIX ‘07 Pin Lu & Kai Shen Department of Computer Science University of Rochester.
Latency Reduction Techniques for Remote Memory Access in ANEMONE Mark Lewandowski Department of Computer Science Florida State University.
VMWare MMU Ranjit Kolkar. Designed for efficient use of resources. ESX uses high-level resource management policies to compute a target memory allocation.
MEMORY RESOURCE MANAGEMENT IN VMWARE ESX SERVER 김정수
Project Presentation By: Dean Morrison 12/6/2006 Dynamically Adaptive Prepaging for Effective Virtual Memory Management.
U NIVERSITY OF M ASSACHUSETTS, A MHERST – Department of Computer Science Dynamic Resource Management in Internet Data Centers Bhuvan Urgaonkar Laboratory.
Full and Para Virtualization
Trusted Passages: Managing Trust Properties of Open Distributed Overlays Faculty: Mustaque Ahamad, Greg Eisenhauer, Wenke Lee and Karsten Schwan PhD Students:
Technical Reading Report Virtual Power: Coordinated Power Management in Virtualized Enterprise Environment Paper by: Ripal Nathuji & Karsten Schwan from.
Cloud Computing – UNIT - II. VIRTUALIZATION Virtualization Hiding the reality The mantra of smart computing is to intelligently hide the reality Binary->
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
Memory Resource Management in VMware ESX Server By Carl A. Waldspurger Presented by Clyde Byrd III (some slides adapted from C. Waldspurger) EECS 582 –
Capacity Planning in a Virtual Environment Chris Chesley, Sr. Systems Engineer
Dynamic Resource Allocation for Shared Data Centers Using Online Measurements By- Abhishek Chandra, Weibo Gong and Prashant Shenoy.
BUFFALO: Bloom Filter Forwarding Architecture for Large Organizations Minlan Yu Princeton University Joint work with Alex Fabrikant,
Virtualization for Cloud Computing
Optimizing Distributed Actor Systems for Dynamic Interactive Services
Presented by Yoon-Soo Lee
Exploiting Sharing for Data Center Consolidation
Zhen Xiao, Qi Chen, and Haipeng Luo May 2013
Specialized Cloud Architectures
Presentation transcript:

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Server Consolidation in Virtualized Data Centers Prashant Shenoy University of Massachusetts

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Networked Applications Proliferation of web-enabled and networked applications Increased use in consumer and business worlds Brokerage/ banking online gameonline store Growing significance in personal, business affairs Focus: networked server applications

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Dynamic Application Workloads Networked apps see dynamic workloads Multi-time-scale variations Time-of-day, hour-of-day Transient spikes Flash crowds Incremental growth User threshold for response time: 8-10 s Key issue: Provide good response time under varying workloads Time (hrs) Request Rate (req/min) Time (hours) Time (days) Arrivals per min K 1200

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Data Centers Networked apps run on data centers Data Centers Large clusters of servers Networked storage devices Allocate resources to meet application SLAs Energy costs are large part of operating budget Modern data centers are increasingly virtualized

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Virtualized Data Centers Virtualized data centers Each application runs inside a virtual server One or more VS mapped onto each physical server Application isolation Dynamic resource allocation VM migration Server consolidation

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Data Center Resource Allocation Growing complexity Static allocation Not suitable for dynamic workloads Resource over-provisioning Resource wastage Estimating peak workloads is hard Manual reallocation Slow allocation time Challenge: How to handle dynamic workloads while efficiently utilizing data center resources? WC Soccer 1998

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Server Consolidation Power is the largest cost for modern data centers Must get as much usage from each server as possible Reduce management costs Virtualization promises great opportunities for consolidation Easily run multiple virtual servers on each host Number of processor cores per server is increasing Memory becomes bottleneck for packing VMs Each VM gets a hard allocation of memory Not as flexible as CPU time How can we fit more VMs per server?

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Our Approach: Exploiting Sharing Content based page sharing If two VMs have identical pages, only store one copy Reduces the total memory required Supported by both Xen and VMware Applications with high sharing potential Anything on Windows Thin client servers Replicated apps Sharing in web applications Operating system files Application libraries ScenarioTotal RAM% Shared 10 WinNT Linux Linux Place similar VMs together to decrease the total memory footprint Source: Memory Resource Management in VMWare ESX Server Waldspurger, OSDI 2002

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Consolidation: Challenges Which VMs are similar? Where should each VM be placed? How to respond to memory hotspots?

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Talk Outline Motivation Memory Buddies architecture VM Fingerprinting Server Consolidation and Hotspot Mitigation Implementation and evaluation Other research projects

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Memory Buddies Architecture Control Plane Sharing Estimation: Determines which VMs have similar memory images Consolidation: uses sharing potential to guide VM placement Hotspot Mitigation: Detects and resolves memory pressure Memory Tracer Tracks memory utilization Creates VM fingerprint

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Memory Tracer Tracks memory contents to estimate sharing Finds list of actively used pages to determine RAM allocation Monitor access bits on each memory page If a bit is set, clear it Later, if bit has been set again, page is in active use Can determine importance of pages in LRU order Use this to predict how much memory is needed to meet target page fault rate

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Content Based Page Sharing Detect identical pages used by multiple virtual machines Calculate a hash for each page for quick comparison Only store one copy of each identical page Use copy-on-write mapping to ensure correctness VM 1 Memory Contents VM 2 Memory Contents Self-sharing Cross VM Sharing Host 1 Host 2 VM 3 Memory Contents

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science VM Fingerprinting Need efficient method to compare VMs’ memory contents Maintaining full hash lists is expensive 32 bit hash/page needs 1MB hash list per 1GB of RAM Use Bloom Filter based fingerprint Probabilistic data structure for storing keys Page hashes are the keys inserted into Bloom Filter Bloom Filter benefits Reduces storage requirement Much faster to compare ? VM 1 Memory VM 2 Memory ?

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Bloom Filter Overview Data structure composed of an array of bits and several hashing functions Insert elements by setting the bit positions resolved from the hashing functions Can lookup an element by checking if those bits are set Hash collisions can lead to false positives Rate depends on array size and number of hash functions

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Bloom Filter Comparison Also useful for fast set intersection Create a Bloom Filter for each set Compare the bits set in each filter’s bit vector Intuition: VM 1 and VM 2 are the most similar since they have more set bits in common Magnitude of sharing can be estimated based on probability of having certain bits set in both vectors z1, z2 = number of zeroes in each bit vector z12 = number of zeroes in AND of vectors m = bit vector size k=number of hash functions VM 1 VM 2 VM 3 Bit Vector

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Server Consolidation Use sharing potential to guide VM placement Step 1: Identify servers to consolidate Servers with low utilization over previous monitoring window Step 2: Determine target hosts Find new candidate host for each VM to be consolidated Place a VM on the host which allows the maximum sharing Also must satisfy CPU, memory, and network constraints Step 3: Perform migrations Use live migration transparent to applications Power down consolidated server after all VMs are moved

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Hotspot Mitigation Memory requirements can change over time Workload or application phase changes increase need Modifying shared pages causes copy-on-write event Step 1: Monitoring system detects hotspot formation Use disk paging and sharing statistics Step 2: Determine cause of hotspot Type 1: swapping occurs without change to sharing rate Type 2: swapping occurs and sharing rate decreases Step 3: Hotspot resolution Type 1: Increase memory allocation if available Type 2: Attempt to migrate VM to a host with higher sharing

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Talk Outline Motivation Memory Buddies architecture VM Fingerprinting Server Consolidation and Hotspot Mitigation Implementation and evaluation Other research projects

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Implementation Uses VMware ESX Server Xen currently lacks simultaneous support for migration and page sharing Control plane Communicates with VMware’s web services based API Memory Tracer Linux Kernel module Gathers page hashes and access statistics Could be placed within the hypervisor layer Testbed Four ESX hosts plus additional control and client machines

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Evaluation: VM Placement Memory Buddies detects which VMs have similar memory contents and groups them accordingly 4 hosts and 4 applications Sharing Aware placement can support two extra VMs

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Evaluation: Data Center Simulation Simulation of 100 host data center Can fit 8 VMs per host without major sharing Effectiveness of sharing changes with application diversity With very few app types random does just as well With very large number there is little chance for benefit At peak, can fit 33% more VMs by utilizing sharing

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Evaluation: Hotspot Mitigation Scenario Host 1: VM 1 and VM 2 Host 2: VM 3 Initially, 1 and 2 have high rate of sharing At time 10, VM 2 makes writes to shared pages Memory Buddies automatically detects loss of sharing and migrates VM 1 to Host 2 where it has a higher sharing rate

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Evaluation: Fingerprinting Bloom Filter comparison accuracy depends on bit vector size For VM with 384MB RAM Less than 1% error for > 64KB Hash list requires 384KB Size increases linearly with RAM Bloom Filters are much faster to compare In 60 seconds: 125 Hash list comparisons 1700 Bloom Filter comparisons

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Large-scale System Monitoring Scenario: Very large data centers (~ 10K servers) Brokerage firms and banks Complex mapping of applications to servers Need scalable techniques to model, monitor, understand and respond in large systems Adaptive monitoring Sample on global-scales Increase monitoring resolution when needed

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Complex Application Modeling Application Models Answer “what if” scenarios about application performance under hypothetical workloads. Data Mining Application Logs Provide clues about application activities and workload characteristics Workload-to-Utilization Models Predict resource utilization based on incoming workload Workload-to-Workload Models Predict incoming workload at subsequent tiers Model Composition Combine models to predict resource requirements several tiers away from monitor

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Related Work Much work in virtualized data centers Summit speakers represent a good sample! Content-based page sharing [OSDI02] Memory monitoring Jones, Arpaci-Dusseau, and Arpaci-Dusseau in ASPLOS ‘06 Lu and Shen in Usenix 2007 Bloom Filters for set intersection Jain, Dahlin, and Tewari in WebDB 2005 Luo, Qin, Geng, and Luo in SKG 2006

U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Conclusions Exploiting page sharing can reduce the number of servers required in a data center Must monitor memory utilization to detect hotspots caused by changes in sharing rates Joint work with Tim Wood, Gabriel Tarasuk-Levin, Jim Cipar, Peter Desnoyers, Mark Corner, and Emery Berger More at Sandpiper source code available