Managing Cloud Resources: Distributed Rate Limiting Alex C. Snoeren Kevin Webb, Bhanu Chandra Vattikonda, Barath Raghavan, Kashi Vishwanath, Sriram Ramabhadran,

Slides:



Advertisements
Similar presentations
Congestion Control and Fairness Models Nick Feamster CS 4251 Computer Networking II Spring 2008.
Advertisements

Scheduling in Web Server Clusters CS 260 LECTURE 3 From: IBM Technical Report.
Hadi Goudarzi and Massoud Pedram
1 Hardware Support for Isolation Krste Asanovic U.C. Berkeley MURI “DHOSA” Site Visit April 28, 2011.
TELE202 Lecture 8 Congestion control 1 Lecturer Dr Z. Huang Overview ¥Last Lecture »X.25 »Source: chapter 10 ¥This Lecture »Congestion control »Source:
William Stallings Data and Computer Communications 7 th Edition Chapter 13 Congestion in Data Networks.
1 © 2013 Cisco and/or its affiliates. All rights reserved. An Improved Hop-by-hop Interest Shaper for Congestion Control in Named Data Networking Yaogong.
Clayton Sullivan PEER-TO-PEER NETWORKS. INTRODUCTION What is a Peer-To-Peer Network A Peer Application Overlay Network Network Architecture and System.
Barath Raghavan, Kashi Vishwanath, Sriram Ramabhadran, Kenneth Yocum, Alex C. Snoeren Defense: Rejaie Johnson, Xian Yi Teng.
Cloud Control with Distributed Rate Limiting Raghaven et all Presented by: Brian Card CS Fall Kinicki 1.
Cloud Control with Distributed Rate Limiting Barath Raghavan, Kashi Vishwanath, Sriram Ramabhadran, Kenneth Yocum, and Alex C. Snoeren University of California,
Doc.: IEEE /0604r1 Submission May 2014 Slide 1 Modeling and Evaluating Variable Bit rate Video Steaming for ax Date: Authors:
CNDS 2001, Phoenix, AZ Simulating the Smart Market Pricing Scheme on Differentiated- Services Architecture Murat Yuksel and Shivkumar Kalyanaraman Rensselaer.
SEEDING CLOUD-BASED SERVICES: DISTRIBUTED RATE LIMITING (DRL) Kevin Webb, Barath Raghavan, Kashi Vishwanath, Sriram Ramabhadran, Kenneth Yocum, and Alex.
PERSISTENT DROPPING: An Efficient Control of Traffic Aggregates Hani JamjoomKang G. Shin Electrical Engineering & Computer Science UNIVERSITY OF MICHIGAN,
Advanced Computer Networking Congestion Control for High Bandwidth-Delay Product Environments (XCP Algorithm) 1.
SEDA: An Architecture for Well- Conditioned, Scalable Internet Services Matt Welsh, David Culler, and Eric Brewer Computer Science Division University.
SDN and Openflow.
Receiver-driven Layered Multicast S. McCanne, V. Jacobsen and M. Vetterli SIGCOMM 1996.
1 Modeling and Emulation of Internet Paths Pramod Sanaga, Jonathon Duerig, Robert Ricci, Jay Lepreau University of Utah.
Web Caching Schemes1 A Survey of Web Caching Schemes for the Internet Jia Wang.
Congestion Dr. Abdulaziz Almulhem. Almulhem©20012 Congestion It occurs when network resources are becoming scarce High demand Over utilized Offered load.
End-to-End Analysis of Distributed Video-on-Demand Systems Padmavathi Mundur, Robert Simon, and Arun K. Sood IEEE Transactions on Multimedia, February.
Multiple constraints QoS Routing Given: - a (real time) connection request with specified QoS requirements (e.g., Bdw, Delay, Jitter, packet loss, path.
1 A General Auction-Based Architecture for Resource Allocation Weidong Cui, Matthew C. Caesar, and Randy H. Katz EECS, UC Berkeley {wdc, mccaesar,
1 Flexlab: A Realistic, Controlled, and Friendly Environment for Evaluating Networked Systems Jonathon Duerig, Robert Ricci, Junxing Zhang, Daniel Gebhardt,
Self-Similarity in Network Traffic Kevin Henkener 5/29/2002.
1 Emulating AQM from End Hosts Presenters: Syed Zaidi Ivor Rodrigues.
Cloud Control with Distributed Rate Limiting Barath Raghavan, Kashi Vishwanath Sriram Ramabhadran, Kenneth Yocum & Alex C.Snoeren Offence: Alex Kiaie &
Scalable Live Video Streaming to Cooperative Clients Using Time Shifting and Video Patching Meng Guo and Mostafa H. Ammar INFOCOM 2004.
Routers with Small Buffers Yashar Ganjali High Performance Networking Group Stanford University
Collecting Correlated Information from a Sensor Network Micah Adler University of Massachusetts, Amherst.
10th Workshop on Information Technologies and Systems 1 A Comparative Evaluation of Internet Pricing Schemes: Smart Market and Dynamic Capacity Contracting.
Junxian Huang 1 Feng Qian 2 Yihua Guo 1 Yuanyuan Zhou 1 Qiang Xu 1 Z. Morley Mao 1 Subhabrata Sen 2 Oliver Spatscheck 2 1 University of Michigan 2 AT&T.
New Challenges in Cloud Datacenter Monitoring and Management
Practical TDMA for Datacenter Ethernet
Slicing the Onion: Anonymity Using Unreliable Overlays Sachin Katti Jeffrey Cohen & Dina Katabi.
Advanced Network Architecture Research Group 2001/11/149 th International Conference on Network Protocols Scalable Socket Buffer Tuning for High-Performance.
A Cloud is a type of parallel and distributed system consisting of a collection of inter- connected and virtualized computers that are dynamically provisioned.
CSC 336 Data Communications and Networking Lecture 8d: Congestion Control : RSVP Dr. Cheer-Sun Yang Spring 2001.
UDT as an Alternative Transport Protocol for GridFTP Raj Kettimuthu Argonne National Laboratory The University of Chicago.
Super-peer Network. Motivation: Search in P2P Centralised (Napster) Flooding (Gnutella)  Essentially a breadth-first search using TTLs Distributed Hash.
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
Covilhã, 30 June Atílio Gameiro Page 1 The information in this document is provided as is and no guarantee or warranty is given that the information is.
Advanced Network Architecture Research Group 2001/11/74 th Asia-Pacific Symposium on Information and Telecommunication Technologies Design and Implementation.
COP 5611 Operating Systems Spring 2010 Dan C. Marinescu Office: HEC 439 B Office hours: M-Wd 2:00-3:00 PM.
TCP with Variance Control for Multihop IEEE Wireless Networks Jiwei Chen, Mario Gerla, Yeng-zhong Lee.
K-Anycast Routing Schemes for Mobile Ad Hoc Networks 指導老師 : 黃鈴玲 教授 學生 : 李京釜.
VMware vSphere Configuration and Management v6
Deadline-based Resource Management for Information- Centric Networks Somaya Arianfar, Pasi Sarolahti, Jörg Ott Aalto University, Department of Communications.
EE 122: Lecture 15 (Quality of Service) Ion Stoica October 25, 2001.
Jennifer Rexford Fall 2014 (TTh 3:00-4:20 in CS 105) COS 561: Advanced Computer Networks TCP.
Fat Virtual Access Points Taken from Srikanth Kandula.
AMQP, Message Broker Babu Ram Dawadi. overview Why MOM architecture? Messaging broker like RabbitMQ in brief RabbitMQ AMQP – What is it ?
TCP Westwood: Efficient Transport for High-speed wired/wireless Networks 2008.
3/12/2013Computer Engg, IIT(BHU)1 CLOUD COMPUTING-1.
Adaptive Inverse Multiplexing for Wide-Area Wireless Networks Alex C. Snoeren MIT Laboratory for Computer Science IEEE Globecom ’99 Rio de Janeiro, December.
INTRODUCTION TO GRID & CLOUD COMPUTING U. Jhashuva 1 Asst. Professor Dept. of CSE.
Accelerating Peer-to-Peer Networks for Video Streaming
Confluent vs. Splittable Flows
Topics discussed in this section:
Network Load Balancing
TCP Congestion Control
Measuring Service in Multi-Class Networks
Video On Demand.
Congestion Control, Internet transport protocols: udp
Streaming Sensor Data Fjord / Sensor Proxy Multiquery Eddy
EE 122: Quality of Service and Resource Allocation
Cloud Computing Architecture
Modeling and Evaluating Variable Bit rate Video Steaming for ax
Presentation transcript:

Managing Cloud Resources: Distributed Rate Limiting Alex C. Snoeren Kevin Webb, Bhanu Chandra Vattikonda, Barath Raghavan, Kashi Vishwanath, Sriram Ramabhadran, and Kenneth Yocum Building and Programming the Cloud Workshop 13 January 2010

2 l Hosting with a single physical presence u However, clients are across the Internet Centralized Internet services Mysore-Park Cloud Workshop – 13 January 2010

3 Cloud-based services l Resources and clients distributed across the world u Often incorporates resources from multiple providers Windows Live Mysore-Park Cloud Workshop – 13 January 2010

4 Resources in the Cloud l Distributed resource consumption u Clients consume resources at multiple sites u Metered billing is state-of-the-art u Service “punished” for popularity »Those unable to pay are disconnected u No control of resources used to serve increased demand l Overprovision and pray u Application designers typically cannot describe needs u Individual service bottlenecks varied but severe »IOps, network bandwidth, CPU, RAM, etc. »Need a way to balance resource demand Mysore-Park Cloud Workshop – 13 January 2010

Two lynchpins for success l Need a way to control and manage distributed resources as if they were centralized u All current models from OS scheduling and provisioning literature assume full knowledge and absolute control u (This talk focuses specifically on network bandwidth) l Must be able to efficiently support rapidly evolving application demand u Balance the resource needs to hardware realization automatically without application designer input u (Another talk if you’re interested) Mysore-Park Cloud Workshop – 13 January 20105

6 S S S D D D 0 ms Limiters Ideal: Emulate a single limiter l Make distributed feel centralized u Packets should experience the same limiter behavior Mysore-Park Cloud Workshop – 13 January 2010

7 Accuracy (how close to K Mbps is delivered, flow rate fairness) + Responsiveness (how quickly demand shifts are accommodated) Vs. Communication Efficiency (how much and often rate limiters must communicate) Engineering tradeoffs Mysore-Park Cloud Workshop – 13 January 2010

8 Limiter 1 Limiter 2 Limiter 3 Limiter 4 Gossip Estimate local demand Estimate interval timer Set allocation Global demand Enforce limit Packet arrival An initial architecture Mysore-Park Cloud Workshop – 13 January 2010

9 Token bucket, fill rate K Mbps Packet Token bucket limiters Mysore-Park Cloud Workshop – 13 January 2010

10 Demand info (bytes/sec) Limiter 1Limiter 2 A global token bucket (GTB)? Mysore-Park Cloud Workshop – 13 January 2010

11 Limiter 1 3 TCP flows S D Limiter 2 7 TCP flows SD Single token bucket 10 TCP flows S D A baseline experiment Mysore-Park Cloud Workshop – 13 January 2010

12 Single token bucket Global token bucket 7 TCP flows 3 TCP flows 10 TCP flows Problem: GTB requires near-instantaneous arrival info GTB performance Mysore-Park Cloud Workshop – 13 January 2010

13 5 Mbps (limit) 4 Mbps (global arrival rate) Case 1: Below global limit, forward packet Limiters send, collect global rate info from others Take 2: Global Random Drop Mysore-Park Cloud Workshop – 13 January 2010

14 5 Mbps (limit) 6 Mbps (global arrival rate) Case 2: Above global limit, drop with probability: Excess / Global arrival rate = 1/6 Same at all limiters Global Random Drop (GRD) Mysore-Park Cloud Workshop – 13 January 2010

15 7 TCP flows 3 TCP flows 10 TCP flows Delivers flow behavior similar to a central limiter GRD baseline performance Single token bucket Global token bucket Mysore-Park Cloud Workshop – 13 January 2010

GRD under dynamic arrivals 16Mysore-Park Cloud Workshop – 13 January 2010 (50-ms estimate interval)

17 Limiter 1 3 TCP flows S D Limiter 2 7 TCP flows S D Returning to our baseline Mysore-Park Cloud Workshop – 13 January 2010

18 “3 flows” “7 flows” Goal: Provide inter-flow fairness for TCP flows Local token-bucket enforcement Basic idea: flow counting Limiter 1Limiter 2 Mysore-Park Cloud Workshop – 13 January 2010

19 Local token rate (limit) = 10 Mbps Flow A = 5 Mbps Flow B = 5 Mbps Flow count = 2 flows Estimating TCP demand 1 TCP flow S S Mysore-Park Cloud Workshop – 13 January 2010

FPS under dynamic arrivals 20Mysore-Park Cloud Workshop – 13 January 2010 (500-ms estimate interval)

21 Comparing FPS to GRD l Both are responsive and provide similar utilization l GRD requires accurate estimates of the global rate at all limiters. GRD (50-ms est. int.) FPS (500-ms est. int.) Mysore-Park Cloud Workshop – 13 January 2010

22 Estimating skewed demand Limiter 1 D Limiter 2 3 TCP flows S D 1 TCP flow S S Mysore-Park Cloud Workshop – 13 January 2010

23 Key insight: Use a TCP flow’s rate to infer demand Local token rate (limit) = 10 Mbps Flow A = 8 Mbps Flow B = 2 Mbps Flow count ≠ demand Bottlenecked elsewhere Estimating skewed demand Mysore-Park Cloud Workshop – 13 January 2010

24 Local token rate (limit) = 10 Mbps Flow A = 8 Mbps Flow B = 2 Mbps Bottlenecked elsewhere Estimating skewed demand 10 8 Local Limit Largest Flow’s Rate = = 1.25 flows Mysore-Park Cloud Workshop – 13 January 2010

25 3 flows Limiter 2 10 Mbps x Global limit = 10 Mbps 1.25 flows Limiter 1 Set local token rate = = 2.94 Mbps Global limit x local flow count Total flow count = FPS example Mysore-Park Cloud Workshop – 13 January 2010

FPS bottleneck example Mysore-Park Cloud Workshop – 13 January l Initially 3:7 split between 10 un-bottlenecked flows l At 25s, 7-flow aggregate bottlenecked to 2 Mbps l At 45s, un-bottlenecked flow arrives: 3:1 for 8 Mbps

Real world constraints l Resources spent tracking usage is pure overhead u Efficient implementation (<3% CPU, sample & hold) u Modest communication budget (<1% bandwidth) l Control channel is slow and lossy u Need to extend gossip protocols to tolerate loss u An interesting research problem on its own… l The nodes themselves may fail or partition u In an asynchronous system, you cannot tell the difference u Need to have a mechanism that deals gracefully with both Mysore-Park Cloud Workshop – 13 January

Robust control communication Mysore-Park Cloud Workshop – 13 January l 7 Limiters enforcing 10 Mbps limit l Demand fluctuates every 5 sec between flows l Varying loss on the control channel

Handling partitions Mysore-Park Cloud Workshop – 13 January l Failsafe operation: each disconnected group k/n l Ideally: Bank-o-mat problem (credit/debit scheme) l Challege: group membership with asymmetric partitions

30 5 Mbps Following PlanetLab demand l Apache Web servers on 10 PlanetLab nodes u 5-Mbps aggregate limit u Shift load over time from 10 nodes to 4 Mysore-Park Cloud Workshop – 13 January 2010

31 Demands at 10 apache servers on Planetlab Demand shifts to just 4 nodes Wasted capacity 31Mysore-Park Cloud Workshop – 13 January 2010 Current limiting options

32 Applying FPS on PlanetLab 32Mysore-Park Cloud Workshop – 13 January 2010

Hierarchical limiting Mysore-Park Cloud Workshop – 13 January

A sample use case Mysore-Park Cloud Workshop – 13 January l T 0: A: 5 flows at L1 l T 55: A: 5 flows at L2 l T 110: B: 5 flows at L1 l T 165: B: 5 flows at L2

Worldwide flow join Mysore-Park Cloud Workshop – 13 January l 8 nodes split between UCSD and Polish Telecom u 5 Mbps aggregate limit u A new flow arrives at each limiter every 10 seconds

Worldwide demand shift Mysore-Park Cloud Workshop – 13 January l Same demand-shift experiment as before u At 50 sec, Polish Telecom demand disappears u Reappears at 90 sec.

37 Where to go from here l Need to “let go” of full control, make decisions with only a “cloudy” view of actual resource consumption u Distinguish between what you know and what you don’t know u Operate efficiently when you know you know. u Have failsafe options when you know you don’t. l Moreover, we cannot rely upon application/service designers to understand their resource demands u The system needs to dynamically adjust to shifts u We’ve started to manage the demand equation u We’re now focusing on the supply side: custom-tailored resource provisioning. Mysore-Park Cloud Workshop – 13 January 2010