Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the

Slides:



Advertisements
Similar presentations
Martin Suchara, Ryan Witt, Bartek Wydrowski California Institute of Technology Pasadena, U.S.A. TCP MaxNet Implementation and Experiments on the WAN in.
Advertisements

Cheng Jin David Wei Steven Low FAST TCP: design and experiments.
Restricted Slow-Start for TCP William Allcock 1,2, Sanjay Hegde 3 and Rajkumar Kettimuthu 1,2 1 Argonne National Laboratory 2 The University of Chicago.
Ahmed El-Hassany CISC856: CISC 856 TCP/IP and Upper Layer Protocols Slides adopted from: Injong Rhee, Lisong Xu.
CUBIC : A New TCP-Friendly High-Speed TCP Variant Injong Rhee, Lisong Xu Member, IEEE v 0.2.
CUBIC Qian HE (Steve) CS 577 – Prof. Bob Kinicki.
Advanced Computer Networking Congestion Control for High Bandwidth-Delay Product Environments (XCP Algorithm) 1.
XCP: Congestion Control for High Bandwidth-Delay Product Network Dina Katabi, Mark Handley and Charlie Rohrs Presented by Ao-Jan Su.
Congestion Control on High-Speed Networks
TCP friendlyness: Progress report for task 3.1 Freek Dijkstra Antony Antony, Hans Blom, Cees de Laat University of Amsterdam CERN, Geneva 25 September.
Recent Research in Congestion Control The problem of high bandwidth-delay product connections By Guillaume Marceau Presented for WPI CS577, Advanced Computer.
Cheng Jin David Wei Steven Low FAST TCP: Motivation, Architecture, Algorithms, Performance.
Rice Networks Group Aleksandar Kuzmanovic Edward W. Knightly Rice University R. Les Cottrell SLAC/SCS-Network Monitoring.
1 Characterization and Evaluation of TCP and UDP-based Transport on Real Networks Les Cottrell, Saad Ansari, Parakram Khandpur, Ruchi Gupta, Richard Hughes-Jones,
The Effect of Router Buffer Size on HighSpeed TCP Performance Dhiman Barman Joint work with Georgios Smaragdakis and Ibrahim Matta.
© 2006 Open Grid Forum Interactions Between Networks, Protocols & Applications HPCN-RG Richard Hughes-Jones OGF20, Manchester, May 2007,
Congestion Control for High Bandwidth-delay Product Networks Dina Katabi, Mark Handley, Charlie Rohrs.
Congestion Control for High Bandwidth-Delay Product Environments Dina Katabi Mark Handley Charlie Rohrs.
Experiences in Design and Implementation of a High Performance Transport Protocol Yunhong Gu, Xinwei Hong, and Robert L. Grossman National Center for Data.
1 Robust Transport Protocol for Dynamic High-Speed Networks: enhancing XCP approach Dino M. Lopez Pacheco INRIA RESO/LIP, ENS of Lyon, France Congduc Pham.
Maximizing End-to-End Network Performance Thomas Hacker University of Michigan October 26, 2001.
Masaki Hirabaru Internet Architecture Group GL Meeting March 19, 2004 High Performance Data transfer on High Bandwidth-Delay Product Networks.
Implementing High Speed TCP (aka Sally Floyd’s) Yee-Ting Li & Gareth Fairey 1 st October 2002 DataTAG CERN (Kinda!)
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
1 Using Netflow data for forecasting Les Cottrell SLAC and Fawad Nazir NIIT, Presented at the CHEP06 Meeting, Mumbai India, February
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
Parallel TCP Bill Allcock Argonne National Laboratory.
1 Overview of IEPM-BW - Bandwidth Testing of Bulk Data Transfer Tools Connie Logg & Les Cottrell – SLAC/Stanford University Presented at the Internet 2.
HighSpeed TCP for High Bandwidth-Delay Product Networks Raj Kettimuthu.
Rate Control Rate control tunes the packet sending rate. No more than one packet can be sent during each packet sending period. Additive Increase: Every.
Congestion Control for High Bandwidth-Delay Product Networks D. Katabi (MIT), M. Handley (UCL), C. Rohrs (MIT) – SIGCOMM’02 Presented by Cheng.
1 Measurements of Internet performance for NIIT, Pakistan Jan – Feb 2004 PingER From Les Cottrell, SLAC For presentation by Prof. Arshad Ali, NIIT.
Masaki Hirabaru NICT Koganei 3rd e-VLBI Workshop October 6, 2004 Makuhari, Japan Performance Measurement on Large Bandwidth-Delay Product.
1 Characterization and Evaluation of TCP and UDP-based Transport on Real Networks Les Cottrell, Saad Ansari, Parakram Khandpur, Ruchi Gupta, Richard Hughes-Jones,
TCP transfers over high latency/bandwidth networks Internet2 Member Meeting HENP working group session April 9-11, 2003, Arlington T. Kelly, University.
Hamilton Institute Evaluating TCP Congestion Control Doug Leith Hamilton Institute Ireland Thanks: Robert Shorten, Yee Ting Lee, Baruch Even.
GNEW2004 CERN March 2004 R. Hughes-Jones Manchester 1 Lessons Learned in Grid Networking or How do we get end-2-end performance to Real Users ? Richard.
Murari Sridharan Windows TCP/IP Networking, Microsoft Corp. (Collaborators: Kun Tan, Jingmin Song, MSRA & Qian Zhang, HKUST)
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
XCP: eXplicit Control Protocol Dina Katabi MIT Lab for Computer Science
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
1 Evaluation of Advanced TCP stacks on Fast Long-Distance production Networks Prepared by Les Cottrell & Hadrien Bullot, Richard Hughes-Jones EPFL, SLAC.
INDIANAUNIVERSITYINDIANAUNIVERSITY Status of FAST TCP and other TCP alternatives John Hicks TransPAC HPCC Engineer Indiana University APAN Meeting – Hawaii.
An Analysis of AIMD Algorithm with Decreasing Increases Yunhong Gu, Xinwei Hong, and Robert L. Grossman National Center for Data Mining.
1 IEPM / PingER project & PPDG Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99 Partially funded by DOE/MICS Field Work Proposal on.
1 FAST TCP for Multi-Gbps WAN: Experiments and Applications Les Cottrell & Fabrizio Coccetti– SLAC Prepared for the Internet2, Washington, April 2003
Network-aware OS DOE/MICS ORNL site visit January 8, 2004 ORNL team: Tom Dunigan, Nagi Rao, Florence Fowler, Steven Carter Matt Mathis Brian.
1 Stochastic Ordering for Internet Congestion Control Han Cai, Do Young Eun, Sangtae Ha, Injong Rhee, and Lisong Xu PFLDnet 2007 February 7, 2007.
Congestion Control for High Bandwidth-Delay Product Networks Dina Katabi, Mark Handley, Charlie Rohrs Presented by Yufei Chen.
© 2006 Andreas Haeberlen, MPI-SWS 1 Monarch: A Tool to Emulate Transport Protocol Flows over the Internet at Large Andreas Haeberlen MPI-SWS / Rice University.
CUBIC Marcos Vieira.
Fast Pattern-Based Throughput Prediction for TCP Bulk Transfers
R. Hughes-Jones Manchester
Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the
TCP-in-UDP draft-welzl-irtf-iccrg-tcp-in-udp-00.txt
TransPAC HPCC Engineer
High Speed File Replication
Using Netflow data for forecasting
HighSpeed TCP for Large Congestion Windows
MB-NG Review High Performance Network Demonstration 21 April 2004
Wide Area Networking at SLAC, Feb ‘03
Characterization and Evaluation of TCP and UDP-based Transport on Real Networks Les Cottrell, Saad Ansari, Parakram Khandpur, Ruchi Gupta, Richard Hughes-Jones,
My Experiences, results and remarks to TCP BW and CT Measurements Tools Jiří Navrátil SLAC.
Fast TCP Matt Weaver CS622 Fall 2007.
Advanced Networking Collaborations at SLAC
Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the
Transport Layer: Congestion Control
An Empirical Evaluation of Wide-Area Internet Bottlenecks
Review of Internet Protocols Transport Layer
Presentation transcript:

Evaluation of Advanced TCP stacks on Fast Long-Distance production Networks Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the BaBar/SCS collaboration November 2003 Partially funded by DOE/MICS Field Work Proposal on Internet End-to-end Performance Monitoring (IEPM), also supported by IUPAP

Project goals Test new advanced TCP stacks, see how they perform on short and long-distance real production WAN links Compare & contrast: ease of configuration, throughput, convergence, fairness, stability etc. For different RTTs, windows, txqueuelen Recommend “optimum” stacks for data intensive science (BaBar) transfers using bbftp, bbcp, GridFTP Validate simulator & emulator findings & provide feedback Problems with real networks: They do not stay still, routes change, links are upgraded, pings are blocked, hosts go down, hosts have configurations changed, routers are misconfigured and/or crash Must avoid impacting production traffic, need agreement of remote administrators and/or network folks Many of the stacks are alpha or beta releases and are in the kernel: kernel crashes, developer wants to add new feature, measurer wants stability or throw out previous measurements. Some stacks were not available in time (Westwood+, GridDT)

Protocol selection TCP only No Rate based transport protocols (e.g. SABUL, UDT, RBUDP) at the moment No iSCSI or FC over IP Sender mods only, HENP model is few big senders, lots of smaller receivers Simplifies deployment, only a few hosts at a few sending sites No DRS Runs on production nets No router mods (XCP/ECN), no jumbos,

Protocols Evaluated Linux 2.4 New Reno with SACK: single and parallel streams (P-TCP) Scalable TCP (S-TCP) Fast TCP HighSpeed TCP (HS-TCP) HighSpeed TCP Low Priority (HSTCP-LP) Binary Increase Control TCP (Bic-TCP) Hamilton TCP (H-TCP)

Reno single stream Low performance on fast long distance paths 700 AIMD (add a=1 pkt to cwnd / RTT, decrease cwnd by factor b=0.5 in congestion) 700 SLAC to Florida Throughput Mbps Reno RTT ms RTT (~70ms) 1200 s

P-TCP TCP Reno with 16 streams Parallel streams heavily used in HENP & elsewhere to achieve needed performance, so it is today’s de facto baseline However, hard to optimize both the window size AND number of streams since optimal values can vary due to network capacity, routes or utilization changes SLAC Caltech May – Nov ’03, note weekend/weekday changes & upgrades

Measurements 20 minute tests, long enough to see stable patterns Iperf reports incremental and cumulative throughputs at 5 second intervals Ping interval about 100ms Exhaustive testing: over a thousand 20 minute measurements, or 300 hours UDP or TCP cross-traffic Xs Xr bottleneck Remote site TCP TCPs TCPr ping ICMP/ping traffic SLAC Ping traffic goes to TCPr when also running cross-traffic Otherwise goes to Xr

Networks 3 main network paths Short distance: SLAC-Caltech (RTT~10ms) Middle distance: UFL and DataTAG Chicago (RTT~70ms) Long distance: CERN and University of Manchester (RTT ~ 170ms) Tests during nights and weekends to avoid unacceptable impacts on production traffic

Windows Set large maximum windows (typically 32MB) on all hosts Used 3 different windows with iperf: Small window size, factor 2-4 below optimal Roughly optimal window size (~BDP) Oversized window

RTT Only P-TCP appears to dramatically affect the RTT E.g. increases by RTT by 200ms (factor 20 for short distances) 700 600 700 600 SLAC-Caltech P-TCP 16 stream SLAC-Caltech FAST TCP 1 stream RTT (ms) RTT (ms) Throughput (Mbps) RTT Throughput (Mbps) RTT 1200 1200 Time (secs) Time (secs)

Throughput SLAC to Remote Throughput (Mbps) Windows too small (worse for longer distance) Throughput SLAC to Remote Reno 16 Sc Bic Fast HS LP H HS Reno 1 Avg Caltech 256 KB 395 226 238 233 236 225 239 253 UFl 1 MB 451 110 133 136 141 140 129 172 Caltech 512 KB 413 377 372 408 374 339 307 362 369 UFl 4 MB 428 437 387 340 383 348 431 294 381 Caltech 1 MB 434 429 382 284 384 UFl 8 MB 442 404 357 351 278 Average 427 327 319 313 312 298 295 279 321 Rank 1 2 4   Poor performance Reasonable performance Best performance Reno with 1 stream has problems on Medium distance link (70ms)

Throughput Avg throughput for optimal & large window sizes from SLAC to CalTech, UFl & Manchester

Cross TCP Traffic Important to understand how fair a protocol is For one protocol competing against the same protocol (intra-protocol) we define the fairness for a single bottleneck as: All protocols have good intra-protocol Fairness (F>0.98) Except HS-TCP (F<0.94) when the window size > optimal 700 600 700 600 SLAC-Caltech Fast-TCP (F~0.997) SLAC-Florida HS-TCP (F~0.935) Aggregate Need good intra-protocol for Florida Aggregate RTT (ms) Throughput (Mbps) Throughput (Mbps) RTT (ms) TCP TCPs RTT RTT 1200 Time (secs) Time (secs) 1200

Inter protocol Fairness For inter-protocol fairness we introduce the asymmetry between the two throughputs: Where x1 and x2 are the throughput averages of TCP stack 1 competing with TCP stack 2 Avg. Asymmetry vs all stacks Avg. Asymmetry vs all stacks Need more granularity of all the asymmetries rather than one average for each, need x-bar in ppt Reno 16 v. aggressive at short RTT, Reno & Scalable aggressive at medium distance HSTCP-LP very timid on medium RTT, HS-TCP also timid

Preliminary Conclusions Advanced stacks behave like TCP-Reno single stream on short distances for up to Gbits/s paths, especially if window size limited TCP Reno single stream has low performance and is unstable on long distances P-TCP is very aggressive and impacts the RTT badly New TCP stacks work well with a single stream, most are fair and stable Ready to try with production applications Heavy reverse traffic may not often be a problem in today’s Internet but it could be more important in future.

More Information TCP Stacks Evaluation: www-iepm.slac.stanford.edu/bw/tcp-eval/