Immediate ECN Mirja Kuhlewind, David Wagner, Juan Manuel Reyes Espinosa, Stuttgart Uni Bob Briscoe, BT IETF-88 TSVAREA Nov 2013 Bob Briscoe was part-funded.

Slides:



Advertisements
Similar presentations
B 黃冠智.
Advertisements

Lecture 18: Congestion Control in Data Center Networks 1.
Mohammad Alizadeh, Albert Greenberg, David A. Maltz, Jitendra Padhye Parveen Patel, Balaji Prabhakar, Sudipta Sengupta, Murari Sridharan Presented by Shaddi.
CSIT560 Internet Infrastructure: Switches and Routers Active Queue Management Presented By: Gary Po, Henry Hui and Kenny Chong.
A Test To Allow TCP Senders to Identify Receiver Non-Compliance Toby Moncaster †, Bob Briscoe*, Arnaud Jacquet* † University of Cambridge * BT draft-moncaster-tcpm-rcv-cheat-03.
PFabric: Minimal Near-Optimal Datacenter Transport Mohammad Alizadeh Shuang Yang, Milad Sharif, Sachin Katti, Nick McKeown, Balaji Prabhakar, Scott Shenker.
Congestion Control: TCP & DC-TCP Swarun Kumar With Slides From: Prof. Katabi, Alizadeh et al.
Balajee Vamanan et al. Deadline-Aware Datacenter TCP (D 2 TCP) Balajee Vamanan, Jahangir Hasan, and T. N. Vijaykumar.
Ahmed Mansy, Mostafa Ammar (Georgia Tech) Bill Ver Steeg (Cisco)
5/17/20151 Adaptive RED: An Algorithm for Increasing the Robustness of RED’s Active Queue Management or How I learned to stop worrying and love RED Presented.
Md. Manzoor MurshedAdaptive RED with Dynamic Threshold Adjustment CprE599: Creative Component Adaptive RED with Dynamic Threshold Adjustment.
Mohammad Alizadeh Adel Javanmard and Balaji Prabhakar Stanford University Analysis of DCTCP:Analysis of DCTCP: Stability, Convergence, and FairnessStability,
Advanced Computer Networking Congestion Control for High Bandwidth-Delay Product Environments (XCP Algorithm) 1.
TFRC for Voice: the VoIP Variant Sally Floyd, Eddie Kohler. March 2005, presentation to AVT draft-ietf-dccp-tfrc-voip-01.txt.
Explicit Congestion Notification (ECN) RFC 3168 Justin Yackoski DEGAS Networking Group CISC856 – TCP/IP Thanks to Namratha Hundigopal.
IETF87 Berlin DCTCP implementation in FreeBSD
The Tension Between High Video Rate and No Rebuffering Te-Yuan (TY) Huang Stanford University IRTF Open 87 July 30th, 2013 Joint work Prof.
XCP: Congestion Control for High Bandwidth-Delay Product Network Dina Katabi, Mark Handley and Charlie Rohrs Presented by Ao-Jan Su.
Receiver-driven Layered Multicast S. McCanne, V. Jacobsen and M. Vetterli SIGCOMM 1996.
Bertha & M Sadeeq.  Easy to manage the problems  Scalability  Real time and real environment  Free data collection  Cost efficient  DCTCP only covers.
Congestion control in data centers
“On Designing Improved Controllers for AQM Routers Supporting TCP Flows” The PI Controller Presented by Bob Kinicki.
The Power of Explicit Congestion Notification Aleksandar Kuzmanovic Northwestern University
1 Minseok Kwon and Sonia Fahmy Department of Computer Sciences Purdue University {kwonm, All our slides and papers.
Defense: Christopher Francis, Rumou duan Data Center TCP (DCTCP) 1.
1 Internet Networking Spring 2003 Tutorial 11 Explicit Congestion Notification (RFC 3168) Limited Transmit (RFC 3042)
1 Internet Networking Spring 2003 Tutorial 11 Explicit Congestion Notification (RFC 3168)
1 Emulating AQM from End Hosts Presenters: Syed Zaidi Ivor Rodrigues.
1 Spring Semester 2007, Dept. of Computer Science, Technion Internet Networking recitation #8 Explicit Congestion Notification (RFC 3168) Limited Transmit.
Random Early Detection Gateways for Congestion Avoidance
The War Between Mice and Elephants By Liang Guo (Graduate Student) Ibrahim Matta (Professor) Boston University ICNP’2001 Presented By Preeti Phadnis.
Ns Simulation Final presentation Stella Pantofel Igor Berman Michael Halperin
Mohammad Alizadeh, Abdul Kabbani, Tom Edsall,
Mohammad Alizadeh Stanford University Joint with: Abdul Kabbani, Tom Edsall, Balaji Prabhakar, Amin Vahdat, Masato Yasuda HULL: High bandwidth, Ultra Low-Latency.
AQM Recommendation Fred Baker. History At IETF 86, TSVAREA decided to update the recommendation of RFC 2309 to not recommend the use of RED Argument:
Curbing Delays in Datacenters: Need Time to Save Time? Mohammad Alizadeh Sachin Katti, Balaji Prabhakar Insieme Networks Stanford University 1.
Adding Explicit Congestion Notification (ECN) Capability to TCP's SYN/ACK Packets A. Kuzmanovic, A. Mondal, S. Floyd, and K.K. Ramakrishnan draft-ietf-tcpm-ecnsyn-03.txt.
B 李奕德.  Abstract  Intro  ECN in DCTCP  TDCTCP  Performance evaluation  conclusion.
ACN: RED paper1 Random Early Detection Gateways for Congestion Avoidance Sally Floyd and Van Jacobson, IEEE Transactions on Networking, Vol.1, No. 4, (Aug.
1 Lecture 14 High-speed TCP connections Wraparound Keeping the pipeline full Estimating RTT Fairness of TCP congestion control Internet resource allocation.
Congestion Control - Supplementary Slides are adapted on Jean Walrand’s Slides.
Byte and Packet Congestion Notification draft-briscoe-tsvwg-byte-pkt-mark-01.txt draft-briscoe-tsvwg-byte-pkt-mark-01.txt Bob Briscoe, BT & UCL IETF-70.
Byte and Packet Congestion Notification draft-briscoe-tsvwg-byte-pkt-mark-00.txt draft-briscoe-tsvwg-byte-pkt-mark-00.txt Bob Briscoe, BT & UCL IETF-69.
DCTCP & CoDel the Best is the Friend of the Good Bob Briscoe, BT Murari Sridharan, Microsoft IETF-84 TSVAREA Jul 2012 Le mieux est l'ennemi du bien Voltaire.
Tunnelling of Explicit Congestion Notification draft-briscoe-tsvwg-ecn-tunnel-03.txt draft-briscoe-tsvwg-ecn-tunnel-03.txt Bob Briscoe, BT IETF-75 saag.
Copyright 2008 Kenneth M. Chipps Ph.D. Controlling Flow Last Update
A Self-Configuring RED Gateway Wu-chang Feng, Dilip Kandlur, Debanjan Saha, Kang Shin INFOCOM ‘99.
AQM & TCP models Courtesy of Sally Floyd with ICIR Raj Jain with OSU.
We used ns-2 network simulator [5] to evaluate RED-DT and compare its performance to RED [1], FRED [2], LQD [3], and CHOKe [4]. All simulation scenarios.
ConEx Concepts and Abstract Mechanism draft-ietf-conex-abstract-mech-01.txt draft-ietf-conex-abstract-mech-01.txt Matt Mathis, Google Bob Briscoe, BT IETF-80.
TCP continued. Discussion – TCP Throughput TCP will most likely generate the saw tooth type of traffic. – A rough estimate is that the congestion window.
ECN the identifier of a new service model Bob Briscoe, BT Mirja Kuhlewind, David Wagner, Stuttgart Uni Koen De Schepper, Alcatel-Lucent IETF-89 TSVAREA.
The Benefits to Applications of using Explicit Congestion Notification (ECN) draft-welzl-ecn-benefits-00 89th IETF Meeting London, UK 4 March 2014 Michael.
1 Sheer volume and dynamic nature of video stresses network resources PIE: A lightweight latency control to address the buffer problem issue Rong Pan,
1 Three ways to (ab)use Multipath Congestion Control Costin Raiciu University Politehnica of Bucharest.
Data Center TCP (DCTCP)
draft-briscoe-tsvwg-l4s-arch-00 B. Briscoe, K. De Schepper, M. Bagnulo
Data Center TCP (DCTCP)
Internet Networking recitation #9
Topics discussed in this section:
Bob Briscoe Simula Research Laboratory
Router-Assisted Congestion Control
Bob Briscoe, BT Murari Sridharan, Microsoft IETF-84 ConEx Jul 2012
Bob Briscoe Simula Research Laboratory
Microsoft Research Stanford University
Michael Welzl University of Oslo
Internet Networking recitation #10
ECN Experimentation draft-black-ecn-experimentation
Lecture 16, Computer Networks (198:552)
Adaptive RED: An Algorithm for Increasing the Robustness of RED’s Active Queue Management or How I learned to stop worrying and love RED Presented by:
Presentation transcript:

Immediate ECN Mirja Kuhlewind, David Wagner, Juan Manuel Reyes Espinosa, Stuttgart Uni Bob Briscoe, BT IETF-88 TSVAREA Nov 2013 Bob Briscoe was part-funded by the European Community under its Seventh Framework Programme through the Reducing Internet Transport Latency (RITE) project ICT

summary & context Promising early results towards the aim of: – the predictably low queuing delay of DCTCP* – deployable on the public Internet, with existing hardware – zero config or config-insensitive At IAB rmcat workshop, we foresaw a need to address: Why bring such early results to the IETF? – to test the water on a redefinition of ECN to foster ECN deployment through more significant benefits problemIETF wg formedthis proposal real-time media congestion avoidancermcat- prevent TCP bloating queuesaqm make TCP smoother- * DCTCP: data centre TCP

problem AQM dynamics buffer’s job: absorb bursts that dissipate by themselves all AQMs defer dropping for c.1 worst-case RTT for a flow with RTT of 20ms or 4ms e.g. content distribution network or home media server these AQMs suppress any signal for 5 or 25 of the flow’s own RTTs CoDel, PIE: auto-tune for varying line-rate also need to auto-tune for varying RTT it’s not ‘good’ to hold back from signalling for 100ms it’s just necessary if the alternative is drop bad queue good(?)queue & 700B/pkt, 512pkt  3s for moving ave of queue to reach 63% of inst queue, but not comparable with PIE & CoDel delays, which are absolute good(?) queue 4ms home media server 20ms CDN 100ms origin

AQM dynamics solution [from DCTCP] For ECN-capable packets – shift the job of smoothing congestion signals from network to host the network signals ECN with no smoothing delay the transport can hide bursts of ECN signals from itself the transport knows whether it's TCP or RTP etc whether its in cong avoidance or slow-start and it knows its own RTT then short RTT flows can smooth the signals themselves delayed only by their own RTT so they can fill troughs and absorb peaks that longer RTT flows cannot so it can decide whether to respond immediately or to smooth the signals and, if so, over what time smoothing congestion signals

aims: real performance gain (and avoid RTT-sensitive config) DCTCP on host uses immediate ECN DCTCP only smooths the ECN signals while in congestion avoidance DCTCP in slow-start responds without smoothing, immediately reducing overshoot * Modified DCTCP is only shown separate from DCTCP, because we improved the original DCTCP slightly

line utilisation buffer occupancy b u f f e r s i z e queue management operating point shallower operating point good line utilisation lower queuing delay buffer kept for bursts TCP saw-teeth seeking the operating point DCTCP: more smaller saw-teeth Today (at best) TCP on end-systems RED in queues if solely change queueschange queues and end-systems cuts delay but poorer line utilisation time Data Centre TCP (DCTCP) high utilisation in steady state still leaves room for bursts highly insensitive to configuration

aim: real performance gain classic ECN cannot justify deployment pain for a questionable performance gain immediate ECN addresses predictability and low queuing delay including self-delay for short flows avoiding RTT-sensitive config

problem II co-existence of DCTCP with existing Internet traffic data centre TCP was so-called only because it couldn’t co-exist with Internet traffic can’t have a low delay threshold for ECN and a deep threshold for drop in one FIFO queue drop traffic would push the queue to its own balance point causing 100% marking of ECN packets then ECN traffic would starve itself ECN drop Queue packet mark/drop probability 0 1 

co-existence solution can use existing network hardware use weighted RED (WRED) implementation in an unusual configuration – one FIFO queue with two instances of RED algo smoothed queue for drop(EWMA-constant = 9 say)* current queue for ECN(EWMA-constant = 0) as share of DCTCP grows – more insensitive to config * if exponential-weighting-constant = B, then RED smooths the queue over 2 B packets if B = 9, RED smooths over 2 9 = 512 packets if B = 0, RED smooths over 2 0 = 1 packet (i.e. it doesn’t smooth) Instantaneous or Averaged Queue q(w q ) ECN drop

a similar coexistence approach should be applicable to other AQMs ultimately, want to auto-tune against line-rate and RTT – use a modern AQM that uses queuing delay as its metric – and separate drop and ECN algos message for implementers in silicon ensure parameters can be configured separately for ECN

co-existence results of ‘gating tests’ explored large part of the much larger parameter space implemented in Linux ; simulated in IKR simlib ‘gating tests’: long-running flows only paper under submission, available on request robust against starvation formula to derive ECN config from drop config to maintain rate fairness can then find sweet spot for the drop config Averaged Queue q(w q ) ECN drop

early deployment, when traffic mostly drop-based have to set drop (and therefore ECN) threshold deep as more flows shift to DCTCP, can set both thresholds shallower early deployment, when traffic mostly drop-based have to set drop (and therefore ECN) threshold deep as more flows shift to DCTCP, can set both thresholds shallower a sample of the results so far Instantaneous queue Averaged queue (w q ) ECN drop

smoothing congestion signals problem III incremental deployment interop between classic and immediate ECN ECN widely implemented on hosts: – on by default at TCP servers – off by default at TCP clients turn clients on by default when deploy: – accurate ECN feedback & ECN fall-back ECN widely implemented on hosts: – on by default at TCP servers – off by default at TCP clients turn clients on by default when deploy: – accurate ECN feedback & ECN fall-back host buffers small smoothed responses to each ECN one big instant response to ECN per RTT immediate ECN 1 smoothed ECN 2 1 don’t get full gain in latency until host upgrades as well 2 doubly delayed response to congestion these two ticks are based on conjecture, not experimental evidence (yet)

cross-layer / cross-wg impact on IETF 1.RFC 3168 may not need to be updated (see spare slide) 2.urgent given pace of AQM development 3.wire protocol: the main standards track change 4.algorithm experimentation expected componentIETF wgdocument 1redefine meaning of ECN CEtsvwgExpt update to RFC3168 2specify ECN behaviour in AQM algosaqmCoDel, PIE, (RED++?) 3specify change to TCP feedbacktcpmdraft-ietf-accurate-ecn-reqs 4specify change to TCP sender algotcpmExpt update to RFC5861

concluding messages research in progress promise of predictably low delay during dynamics an unnecessary queue is not a ‘good’ queue adds RTT auto-tuning to AQM by shifting smoothing from network to host can use existing network hardware if you’re implementing a new AQM at least ensure parameters can be configured separately for ECN question: if subsequent experiments are as promising as these, would there be an appetite in the transport area to tweak the meaning of ECN? research in progress promise of predictably low delay during dynamics an unnecessary queue is not a ‘good’ queue adds RTT auto-tuning to AQM by shifting smoothing from network to host can use existing network hardware if you’re implementing a new AQM at least ensure parameters can be configured separately for ECN question: if subsequent experiments are as promising as these, would there be an appetite in the transport area to tweak the meaning of ECN?

Immediate ECN Q&A spare slides

which codepoint for immediate ECN? To use CE for immediate ECN, may not need to update RFC3168 (Addition of ECN to IP):...if the ECT codepoint is set in that packet's IP header... then instead of dropping the packet, the router MAY instead set the CE codepoint in the IP header. An environment where all end nodes were ECN-Capable could allow new criteria to be developed for setting the CE codepoint, and new congestion control mechanisms for end- node reaction to CE packets. However, this is a research issue, and as such is not addressed in this document. Could use ECT(1) for immediate ECN but this unnecessarily wastes the CE codepoint (who would want ‘sluggish ECN’?)

DCTCP in Action 20 Setup: Win 7, Broadcom 1Gbps Switch Scenario: 2 long-lived flows, K = 30KB (Kbytes)

18 Parameters: link capacity = 10Gbps RTT = 480μs smoothing constant (at source), g = For TCP: Throughput → 75% For TCP: Throughput → 75% Throughput-Latency Tradeoff Throughput > 94% as K  0

DCTCP activity E2e Transport – In Windows 8 Server data center template – I-D for DCTCP feedback (intended EXP) [draft-kuehlewind-tcpm-accurate-ecn-01] AQM – Existing kit: Just a degenerate config of RED – Can be implemented as just a step at K packets (single ‘if’ command) – For zero-delay can use a virtual queue [RC5670] hardware implementations [“How to Build a Virtual Queue from Two Leaky Buckets”]How to Build a Virtual Queue from Two Leaky Buckets see HULL for specifics with DCTCPHULL Analysis, papers, Linux & ns2 implementation, etc – – SIGCOMM paper gives entry point Averaged Figures courtesy of Alizadeh et al

Data Center TCP Algorithm Switch side: Mark packets when Queue Length > K Sender side: Maintain moving average of fraction of marked packets (α) Adaptive congestion window decrease: B K Mark Don’t Mark