TCP Performance over a 2.5 Gbit/s Transatlantic Circuit

Slides:



Advertisements
Similar presentations
The DataTAG Project 25 March, Brussels FP6 Information Day Peter Clarke, University College London.
Advertisements

TCP transfers over high latency/bandwidth network & Grid TCP Sylvain Ravot
Cheng Jin David Wei Steven Low FAST TCP: design and experiments.
Ahmed El-Hassany CISC856: CISC 856 TCP/IP and Upper Layer Protocols Slides adopted from: Injong Rhee, Lisong Xu.
Presentation by Joe Szymanski For Upper Layer Protocols May 18, 2015.
School of Information Technologies TCP Congestion Control NETS3303/3603 Week 9.
Congestion Control on High-Speed Networks
Efficient Network Protocols for Data-Intensive Worldwide Grids Seminar at JAIST, Japan 3 March 2003 T. Kelly, University of Cambridge, UK S. Ravot, Caltech,
Congestion Control Tanenbaum 5.3, /12/2015Congestion Control (A Loss Based Technique: TCP)2 What? Why? Congestion occurs when –there is no reservation.
02 nd April 03Networkshop Managed Bandwidth Next Generation F. Saka UCL NETSYS (NETwork SYStems centre of excellence)
KEK Network Qi Fazhi KEK SW L2/L3 Switch for outside connections Central L2/L3 Switch A Netscreen Firewall Super Sinet Router 10GbE 2 x GbE IDS.
UDT: UDP based Data Transfer Yunhong Gu & Robert Grossman Laboratory for Advanced Computing University of Illinois at Chicago.
NORDUnet 2003, Reykjavik, Iceland, 26 August 2003 High-Performance Transport Protocols for Data-Intensive World-Wide Grids T. Kelly, University of Cambridge,
MAIN TECHNICAL CHARACTERISTICS Next generation optical transport networks with 40Gbps capabilities are expected to be based on the ITU’s.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
Experience with Loss-Based Congestion Controlled TCP Stacks Yee-Ting Li University College London.
FAST TCP in Linux Cheng Jin David Wei Steven Low California Institute of Technology.
High TCP performance over wide area networks Arlington, VA May 8, 2002 Sylvain Ravot CalTech HENP Working Group.
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
HighSpeed TCP for High Bandwidth-Delay Product Networks Raj Kettimuthu.
23 January 2003Paolo Moroni (Slide 1) SWITCH-cc meeting DataTAG overview.
BIC Control for Fast Long-Distance Networks paper written by Injong Rhee, Lisong Xu & Khaled Harfoush (2004) Presented by Jonathan di Costanzo (2009/02/18)
Data Transport Challenges for e-VLBI Julianne S.O. Sansa* * With Arpad Szomoru, Thijs van der Hulst & Mike Garret.
Internet data transfer record between CERN and California Sylvain Ravot (Caltech) Paolo Moroni (CERN)
TERENA Networking Conference, Zagreb, Croatia, 21 May 2003 High-Performance Data Transport for Grid Applications T. Kelly, University of Cambridge, UK.
Transport Layer 3-1 Chapter 3 Transport Layer Computer Networking: A Top Down Approach 6 th edition Jim Kurose, Keith Ross Addison-Wesley March
Tiziana FerrariThe DataTAG Projct, Roma Nov DataTAG Project.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract The main idea behind the DataTAG project was to strengthen the collaboration.
TCP transfers over high latency/bandwidth networks Internet2 Member Meeting HENP working group session April 9-11, 2003, Arlington T. Kelly, University.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
30 June Wide Area Networking Performance Challenges Olivier Martin, CERN UK DTI visit.
DataTAG: Glue schema test C.Vistoli DataTAG Wp4 Meeting 23/5/2002.
Data Transport Challenges for e-VLBI Julianne S.O. Sansa* * With Arpad Szomoru, Thijs van der Hulst & Mike Garret.
18/09/2002Presentation to Spirent1 Presentation to Spirent 18/09/2002.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
INDIANAUNIVERSITYINDIANAUNIVERSITY Status of FAST TCP and other TCP alternatives John Hicks TransPAC HPCC Engineer Indiana University APAN Meeting – Hawaii.
1 Testing TCP Westwood+ over Transatlantic Links at 10 Gigabit/Second rate Saverio Mascolo Dipartimento di Elettrotecnica ed Elettronica Politecnico di.
Peer-to-Peer Networks 13 Internet – The Underlay Network
The EU DataTAG Project Richard Hughes-Jones Based on Olivier H. Martin GGF3 Frascati, Italy Oct 2001.
DataTAG overview. 3 February 2003Paolo Moroni (Slide 2) APM meeting - Barcelona Summary  Why DataTAG?  DataTAG project  Test-bed extensions  General.
1 FAST TCP for Multi-Gbps WAN: Experiments and Applications Les Cottrell & Fabrizio Coccetti– SLAC Prepared for the Internet2, Washington, April 2003
Network-aware OS DOE/MICS ORNL site visit January 8, 2004 ORNL team: Tom Dunigan, Nagi Rao, Florence Fowler, Steven Carter Matt Mathis Brian.
Wide Area Network Performance Analysis Methodology
TESTBED (Technical)
R. Hughes-Jones Manchester
Networking between China and Europe
Networking for grid Network capacity Network throughput
TransPAC HPCC Engineer
CERN-USA connectivity update DataTAG project
The DataTAG Project Olivier H. Martin
Internet Congestion Control Research Group
CEOS workshop on Grid computing Slides mainly from Olivier Martin
Lecture 19 – TCP Performance
DataTAG Project update
MB-NG Review High Performance Network Demonstration 21 April 2004
Wide Area Networking at SLAC, Feb ‘03
Presented at the GGF3 conference 8th October Frascati, Italy
The EU DataTAG Project Olivier H. Martin CERN - IT Division
So far, On the networking side, we looked at mechanisms to links hosts using direct linked networks and then forming a network of these networks. We introduced.
5th EU DataGrid Conference
Presentation at University of Twente, The Netherlands
The DataTAG Project UCSD/La Jolla, USA Olivier H. Martin / CERN
Wide-Area Networking at SLAC
Internet2 Spring Member meeting
Presented at the 4th DataGrid Conference
High-Performance Data Transport for Grid Applications
Review of Internet Protocols Transport Layer
Presentation transcript:

TCP Performance over a 2.5 Gbit/s Transatlantic Circuit HNF-Europe Workshop 2002, ECMWF, UK 11 October 2002 J.P. Martin-Flatin, CERN S. Ravot, Caltech

The DataTAG Project http://www.datatag.org/

J.P. Martin-Flatin and S. Ravot Project Partners EU-funded partners: CERN (CH), INFN (IT), INRIA (FR), PPARC (UK) and University of Amsterdam (NL) U.S.-funded partners: Caltech, UIC, UMich, Northwestern University, StarLight Collaborations: GEANT, Internet2, Abilene, Canarie, SLAC, ANL, FNAL, LBNL, etc. Project coordinator: CERN 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot About DataTAG Budget: EUR 3.98M Funded manpower: 15 FTE/year 21 people recruited 26 people externally funded Start date: January 1, 2002 Duration: 2 years 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot Three Objectives Build a testbed to experiment with massive file transfers across the Atlantic High-performance protocols for gigabit networks underlying data-intensive Grids Interoperability between several major Grid projects in Europe and USA 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot GIIS giis.ivdgl.org mds-vo-name=glue Gatekeeper: Padova-site Grids GIIS edt004.cnaf.infn.it Mds-vo-name=‘Datatag’ LSF Resource Broker Gatekeeper: US-CMS GIIS giis.ivdgl.org mds-vo-name=ivdgl-glue Gatekeeper grid006f.cnaf.infn.it Condor Gatekeeper edt004.cnaf.infn.it Gatekeeper: US-ATLAS WN1 edt001.cnaf.infn.it WN2 edt002.cnaf.infn.it Computing Element-1 PBS Computing Element -2 Fork/pbs Gatekeeper LSF hamachi.cs.uchicago.edu dc-user.isi.edu rod.mcs.anl.gov Job manager: Fork DataTAG iVDGL 11 October 2002 J.P. Martin-Flatin and S. Ravot

Testbed

J.P. Martin-Flatin and S. Ravot Objectives Provisioning of 2.5 Gbit/s transatlantic circuit between CERN (Geneva) and StarLight (Chicago) Dedicated to research (no production traffic) Multi-vendor testbed with layer-2 and layer-3 capabilities: Cisco Juniper Alcatel Extreme Networks Testbed open to other Grid projects 11 October 2002 J.P. Martin-Flatin and S. Ravot

2.5 Gbit/s Transatlantic Circuit Operational since 20 August 2002 (T-Systems) Circuit initially connected to Cisco 7606 routers (layer 3) High-end PC servers at CERN and StarLight: 6x SuperMicro 2x2.4 GHz SysKonnect SK-9843 GbE cards (2 per PC) can saturate the circuit with TCP traffic ready for upgrade to 10 Gbit/s Deployment of layer-2 equipment under way Testbed fully deployed by 31 October 2002 11 October 2002 J.P. Martin-Flatin and S. Ravot

R&D Connectivity Between Europe & USA NL SURFnet CH CERN UK SuperJANET4 IT GARR-B GEANT FR INRIA FR VTHD ATRIUM 3*2.5Gbit/s Abilene ESNET MREN New York StarLight Canarie

Network Research

Network Research Activities Enhance TCP performance for massive file transfers Monitoring QoS LBE (Scavenger) Bandwidth reservation AAA-based bandwidth on demand lightpath managed as a Grid resource 11 October 2002 J.P. Martin-Flatin and S. Ravot

TCP Performance Issues: The Big Picture TCP’s current congestion control (AIMD) algorithms are not suited to gigabit networks Line errors are interpreted as congestion Delayed ACKs + large cwnd + large RTT = problem 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot AIMD Algorithms Van Jacobson, SIGCOMM 1988 Additive Increase: a TCP connection increases slowly its bandwidth use in the absence of loss forever, unless we run out of send/receive buffers TCP is greedy: no attempt to reach a stationary state Slow start: increase after each ACK Congestion avoidance: increase once per RTT Multiplicative Decrease: a TCP connection reduces its bandwidth use by half after a loss is detected 11 October 2002 J.P. Martin-Flatin and S. Ravot

Congestion Window (cwnd) average cwnd over the last 10 samples average cwnd over the entire lifetime of the connection Slow Start Congestion Avoidance SSTHRESH tcptrace

Disastrous Effect of Packet Loss on TCP in WANs (1/2) TCP throughput as a function of time MSS=1460 11 October 2002 J.P. Martin-Flatin and S. Ravot

Disastrous Effect of Packet Loss on TCP in WANs (2/2) Long time to recover from a single loss: TCP should react to congestion rather than packet loss (line errors, transient fault in equipment) TCP should recover quicker from a loss TCP is much more sensitive to packet loss in WANs than in LANs 11 October 2002 J.P. Martin-Flatin and S. Ravot

Measurements with Different MTUs (1/2) Experimental environment: Linux 2.4.19 Traffic generated by iperf average throughout over the last 5 seconds Single TCP stream RTT = 119 ms Duration of each test: 2 hours Transfers from Chicago to Geneva MTUs: set on the NIC of the PC (ifconfig) POS MTU set to 9180 Max MTU with Linux 2.4.19: 9000 11 October 2002 J.P. Martin-Flatin and S. Ravot

Measurements with Different MTUs (2/2) TCP max: 990 Mbit/s (MTU=9000) UDP max: 957 Mbit/s (MTU=1500) 11 October 2002 J.P. Martin-Flatin and S. Ravot

Tools Used to Perform Measurements We used several tools to investigate TCP performance issues: Generation of TCP flows: iperf and gensink Capture of packet flows: tcpdump tcpdump output=> tcptrace =>xplot Some tests performed with SmartBits 2000 Currently shopping for a traffic analyzer Suggestions? 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot Responsiveness The responsiveness r measures how quickly we go back to using the network link at full capacity after experiencing a loss C . RTT 2 r = 2 . inc 11 October 2002 J.P. Martin-Flatin and S. Ravot

Characterization of the Problem inc size = MSS = 1,460 Capacity RTT # inc Responsiveness 9.6 kbit/s (WAN 1988) max: 40 ms 1 0.6 ms 10 Mbit/s (LAN 1988) max: 20 ms 8 ~150 ms 100 Mbit/s (LAN 2002) max: 5 ms 20 ~100 ms 622 Mbit/s 120 ms ~2,900 ~6 min 2.5 Gbit/s ~11,600 ~23 min 10 Gbit/s ~46,200 ~1h 30min

J.P. Martin-Flatin and S. Ravot What Can We Do? To achieve high throughput over high latency/bandwidth network, we need to: Set the initial slow start threshold (ssthresh) to an appropriate value for the delay and bandwidth of the link Avoid loss by limiting the max size of cwnd Recover fast in case of loss: larger cwnd increment => better responsiveness larger packet size (Jumbo frames) Less aggressive decrease algorithm 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot Delayed ACKs RFC 2581: Assumption: one ACK per packet Delayed ACKs: one ACK every second packet Responsiveness multiplied by two Disastrous effect when RTT large and cwnd large 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot Research Directions New fairness principle Change multiplicative decrease: do not divide by two Change additive increase successive dichotomies local and global stability More stringent definition of congestion Estimation of the available capacity and bandwidth*delay product: on the fly cached 11 October 2002 J.P. Martin-Flatin and S. Ravot

J.P. Martin-Flatin and S. Ravot Related Work Sally Floyd, ICIR: Internet-Draft “High Speed TCP for Large Congestion Windows” Steven Low, Caltech: Fast TCP Tom Kelly, U Cambridge: Scalable TCP Web100 and Net100 PFLDnet 2003 workshop: http://www.datatag.org/pfldnet2003/ 11 October 2002 J.P. Martin-Flatin and S. Ravot