DCTCP & CoDel the Best is the Friend of the Good Bob Briscoe, BT Murari Sridharan, Microsoft IETF-84 TSVAREA Jul 2012 Le mieux est l'ennemi du bien Voltaire.

Slides:



Advertisements
Similar presentations
Mohammad Alizadeh, Albert Greenberg, David A. Maltz, Jitendra Padhye Parveen Patel, Balaji Prabhakar, Sudipta Sengupta, Murari Sridharan Microsoft Research.
Advertisements

B 黃冠智.
Immediate ECN Mirja Kuhlewind, David Wagner, Juan Manuel Reyes Espinosa, Stuttgart Uni Bob Briscoe, BT IETF-88 TSVAREA Nov 2013 Bob Briscoe was part-funded.
Mohammad Alizadeh, Albert Greenberg, David A. Maltz, Jitendra Padhye Parveen Patel, Balaji Prabhakar, Sudipta Sengupta, Murari Sridharan Modified by Feng.
Lecture 18: Congestion Control in Data Center Networks 1.
Mohammad Alizadeh, Albert Greenberg, David A. Maltz, Jitendra Padhye Parveen Patel, Balaji Prabhakar, Sudipta Sengupta, Murari Sridharan Presented by Shaddi.
1 CONGESTION CONTROL. 2 Congestion Control When one part of the subnet (e.g. one or more routers in an area) becomes overloaded, congestion results. Because.
Congestion Control Tanenbaum 5.3 Tanenbaum 6.5. Congestion Control Network Layer – Congestion control point to point Transport Layer – Congestion control.
1.  Congestion Control Congestion Control  Factors that Cause Congestion Factors that Cause Congestion  Congestion Control vs Flow Control Congestion.
PFabric: Minimal Near-Optimal Datacenter Transport Mohammad Alizadeh Shuang Yang, Milad Sharif, Sachin Katti, Nick McKeown, Balaji Prabhakar, Scott Shenker.
Congestion Control: TCP & DC-TCP Swarun Kumar With Slides From: Prof. Katabi, Alizadeh et al.
Balajee Vamanan et al. Deadline-Aware Datacenter TCP (D 2 TCP) Balajee Vamanan, Jahangir Hasan, and T. N. Vijaykumar.
5/17/20151 Adaptive RED: An Algorithm for Increasing the Robustness of RED’s Active Queue Management or How I learned to stop worrying and love RED Presented.
Mohammad Alizadeh Adel Javanmard and Balaji Prabhakar Stanford University Analysis of DCTCP:Analysis of DCTCP: Stability, Convergence, and FairnessStability,
Explicit Congestion Notification (ECN) RFC 3168 Justin Yackoski DEGAS Networking Group CISC856 – TCP/IP Thanks to Namratha Hundigopal.
IETF87 Berlin DCTCP implementation in FreeBSD
XCP: Congestion Control for High Bandwidth-Delay Product Network Dina Katabi, Mark Handley and Charlie Rohrs Presented by Ao-Jan Su.
Bertha & M Sadeeq.  Easy to manage the problems  Scalability  Real time and real environment  Free data collection  Cost efficient  DCTCP only covers.
Congestion control in data centers
The Power of Explicit Congestion Notification Aleksandar Kuzmanovic Northwestern University
Congestion Control Tanenbaum 5.3, /12/2015Congestion Control (A Loss Based Technique: TCP)2 What? Why? Congestion occurs when –there is no reservation.
Explicit Congestion Notification ECN Tilo Hamann Technical University Hamburg-Harburg, Germany.
Defense: Christopher Francis, Rumou duan Data Center TCP (DCTCP) 1.
1 Internet Networking Spring 2003 Tutorial 11 Explicit Congestion Notification (RFC 3168) Limited Transmit (RFC 3042)
High speed TCP’s. Why high-speed TCP? Suppose that the bottleneck bandwidth is 10Gbps and RTT = 200ms. Bandwidth delay product is packets (1500.
1 Minseok Kwon and Sonia Fahmy Department of Computer Sciences Purdue University {kwonm, TCP Increase/Decrease.
1 Internet Networking Spring 2003 Tutorial 11 Explicit Congestion Notification (RFC 3168)
1 Emulating AQM from End Hosts Presenters: Syed Zaidi Ivor Rodrigues.
Data Communication and Networks
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #8 Explicit Congestion Notification (RFC 3168) Limited Transmit.
Mohammad Alizadeh, Abdul Kabbani, Tom Edsall,
ICTCP: Incast Congestion Control for TCP in Data Center Networks∗
Mohammad Alizadeh Stanford University Joint with: Abdul Kabbani, Tom Edsall, Balaji Prabhakar, Amin Vahdat, Masato Yasuda HULL: High bandwidth, Ultra Low-Latency.
AQM Recommendation Fred Baker. History At IETF 86, TSVAREA decided to update the recommendation of RFC 2309 to not recommend the use of RED Argument:
TCP & Data Center Networking
Congestion Control In The Internet JY Le Boudec Fall
Curbing Delays in Datacenters: Need Time to Save Time? Mohammad Alizadeh Sachin Katti, Balaji Prabhakar Insieme Networks Stanford University 1.
Transport Layer3-1 Chapter 3 outline r 3.1 Transport-layer services r 3.2 Multiplexing and demultiplexing r 3.3 Connectionless transport: UDP r 3.4 Principles.
Understanding the Performance of TCP Pacing Amit Aggarwal, Stefan Savage, Thomas Anderson Department of Computer Science and Engineering University of.
Adding Explicit Congestion Notification (ECN) Capability to TCP's SYN/ACK Packets A. Kuzmanovic, A. Mondal, S. Floyd, and K.K. Ramakrishnan draft-ietf-tcpm-ecnsyn-03.txt.
B 李奕德.  Abstract  Intro  ECN in DCTCP  TDCTCP  Performance evaluation  conclusion.
Congestion Control - Supplementary Slides are adapted on Jean Walrand’s Slides.
Congestion control for Multipath TCP (MPTCP) Damon Wischik Costin Raiciu Adam Greenhalgh Mark Handley THE ROYAL SOCIETY.
Byte and Packet Congestion Notification draft-briscoe-tsvwg-byte-pkt-mark-00.txt draft-briscoe-tsvwg-byte-pkt-mark-00.txt Bob Briscoe, BT & UCL IETF-69.
Congestion exposure BoF candidate protocol: re-ECN Bob Briscoe Chief Researcher, BT Nov 2009 This work is partly funded by Trilogy, a research project.
AQM & TCP models Courtesy of Sally Floyd with ICIR Raj Jain with OSU.
TCP: Transmission Control Protocol Part II : Protocol Mechanisms Computer Network System Sirak Kaewjamnong Semester 1st, 2004.
1 SIGCOMM ’ 03 Low-Rate TCP-Targeted Denial of Service Attacks A. Kuzmanovic and E. W. Knightly Rice University Reviewed by Haoyu Song 9/25/2003.
Thoughts on the Evolution of TCP in the Internet (version 2) Sally Floyd ICIR Wednesday Lunch March 17,
TCP continued. Discussion – TCP Throughput TCP will most likely generate the saw tooth type of traffic. – A rough estimate is that the congestion window.
ECN the identifier of a new service model Bob Briscoe, BT Mirja Kuhlewind, David Wagner, Stuttgart Uni Koen De Schepper, Alcatel-Lucent IETF-89 TSVAREA.
Uni Innsbruck Informatik th IETF, PMTUD WG: Path MTU Discovery Using Options draft-welzl-pmtud-options-01.txt Michael Welzl
1 Sheer volume and dynamic nature of video stresses network resources PIE: A lightweight latency control to address the buffer problem issue Rong Pan,
6.888: Lecture 3 Data Center Congestion Control Mohammad Alizadeh Spring
1 Three ways to (ab)use Multipath Congestion Control Costin Raiciu University Politehnica of Bucharest.
Adding Explicit Congestion Notification (ECN) Capability to TCP's SYN/ACK Packets A. Kuzmanovic, A. Mondal, S. Floyd, and K.K. Ramakrishnan draft-ietf-tcpm-ecnsyn-02.txt.
Revisiting Transport Congestion Control Jian He UT Austin 1.
Mohammad Alizadeh, Albert Greenberg, David A. Maltz, Jitendra Padhye Parveen Patel, Balaji Prabhakar, Sudipta Sengupta, Murari Sridharan Microsoft Research.
Data Center TCP (DCTCP)
draft-briscoe-tsvwg-l4s-arch-00 B. Briscoe, K. De Schepper, M. Bagnulo
Data Center TCP (DCTCP)
Internet Networking recitation #9
On Queuing, Marking, and Dropping
Topics discussed in this section:
Bob Briscoe Simula Research Laboratory
CONGESTION CONTROL.
Microsoft Research Stanford University
Internet Networking recitation #10
Lecture 16, Computer Networks (198:552)
Adaptive RED: An Algorithm for Increasing the Robustness of RED’s Active Queue Management or How I learned to stop worrying and love RED Presented by:
Presentation transcript:

DCTCP & CoDel the Best is the Friend of the Good Bob Briscoe, BT Murari Sridharan, Microsoft IETF-84 TSVAREA Jul 2012 Le mieux est l'ennemi du bien Voltaire

menu 1.why DCTCP is important 2.an (untested) roadmap how might DCTCP be deployed and co- exist with current Internet traffic & AQMs

line utilisation buffer utilisation b u f f e r s i z e queue management operating point shallower operating point good line utilisation lower queuing delay buffer kept for bursts TCP saw-teeth seeking the operating point DCTCP: more smaller saw-teeth Today TCP on end-systems RED in queues if solely change queueschange queues and end-systems cuts delay but poorer line utilisation time Data Centre TCP (DCTCP) high utilisation in steady state still leaves room for bursts

Data Center TCP Algorithm Switch side: Mark packets when Queue Length > K Sender side (differences from TCP New Reno): Maintain moving average of fraction of marked packets (α) Adaptive congestion window decrease: B K Mark Don’t Mark

DCTCP in Action 20 Setup: Win 7, Broadcom 1Gbps Switch Scenario: 2 long-lived flows, K = 30KB (Kbytes)

18 Parameters: link capacity = 10Gbps RTT = 480μs smoothing constant (at source), g = For TCP: Throughput → 75% For TCP: Throughput → 75% Throughput-Latency Tradeoff Throughput > 94% as K  0

DCTCP only for data centres? named for a feasible deployment scenario a change to all senders, receivers and switches* not intended to be its sole applicability addresses high bandwidth-delay product should be applicable to slow links & long RTTs only tested down to 100Mb/s so far † * Switches/routers only require reconfig if they support ECN senders (and receivers) require implementation change † An issue with a wide range of RTTs has been addressed 100Mb/s x 500 μs = 250kb/s x 200ms

DCTCP activity E2e Transport In Windows 8 Server data center template I-D for DCTCP feedback (intended EXP) [draft-kuehlewind-tcpm-accurate-ecn-01] AQM Existing kit: Just a degenerate config of RED Can be implemented as just a step at K packets (single ‘if’ command) For zero-delay can use a virtual queue [RC5670] hardware implementations [“How to Build a Virtual Queue from Two Leaky Buckets”]How to Build a Virtual Queue from Two Leaky Buckets see HULL for specifics with DCTCPHULL Analysis, papers, Linux & ns2 implementation, etc SIGCOMM paper gives entry point Averaged Figures courtesy of Alizadeh et al

DCTCP: differences from traditional AQMs e.g. CoDel source smooths signals not the queue source responds to extent of signals not just their existence designed for ECN only

which node owns the RTT? we want to smooth away queues that disappear within ~1 RTT but which RTT? the network? traditional AQMs hold back signals for the ‘nominal’ worst-case (long RTT) DCTCP signals immediately no ‘nominal’ RTT to configure / hard-code / adapt the host? each DCTCP flow smooths over its own RTT short RTT flows can fill troughs & avoid peaks on behalf of longer ones (and themselves) inst. queue length traditional AQM signals are just getting started ~5ms local RTT incl. caches 100ms nominal continental RTT time

DCTCP & CoDel the Best is the Friend of the Good But DCTCP’s approach only makes sense with ECN… Encore, le mieux est l'ennemi du bien

ECN and drop are not equivalent ECN is solely a signal no problem sending out a burst of ECN and later smoothing it away at the source drop is both an impairment and a signal simultaneously want to avoid it and hear it a burst of loss can’t just be smoothed away –collateral damage from timeouts etc.

Can smoothing on the host interoperate with smoothing in the network? current rule (paraphrased from RFC 3168) Signal ECN when queue would otherwise drop Respond to ECN exactly as a drop intended to prevent starvation of one by the other proposal: overload the meaning of an ECN-capable pkt for queue, ECN also means SHOULD NOT smooth for transport, ECN also means SHOULD smooth under persistent congestion need to ensure shares stabilise and no-one starves despite different dynamics

interoperability between old & overloaded meanings of ECN host queue one big instant response to ECN per RTT small smoothed responses to each ECN smoothed ECN 2 instant ECN 1 ticks are based on conjecture, not experimental evidence 1 don’t get full gain in latency until host upgrades as well 2 doubly delayed response to congestion

message zero-config AQMs are good CoDel for drop simple step for ECN –far greater potential gains in parallel to CoDel field testing work on interop with unsmoothed AQM for ECN otherwise the lazy option (ECN = drop) prevails would be a wasted opportunity

DCTCP & CoDel the Best is the Friend of the Good Q&A Je ne savais pas le mieux était si simple

Data Center TCP Algorithm Switch side: Mark packets when Queue Length > K Sender side: Maintain moving average of fraction of marked packets (α) Adaptive congestion window decrease: B K Mark Don’t Mark