Masaki Hirabaru CRL, Japan ITRC MAI BoF Kochi November 6, 2003 広帯域・高遅延ネットワークでの TCP性能計測.

Slides:



Advertisements
Similar presentations
Martin Suchara, Ryan Witt, Bartek Wydrowski California Institute of Technology Pasadena, U.S.A. TCP MaxNet Implementation and Experiments on the WAN in.
Advertisements

Engineering Meeting Report Aug. 29, 2003 Kazunori Konishi.
Helping TCP Work at Gbps Cheng Jin the FAST project at Caltech
Tuning and Evaluating TCP End-to-End Performance in LFN Networks P. Cimbál* Measurement was supported by Sven Ubik**
FAST TCP Anwis Das Ajay Gulati Slides adapted from : IETF presentation slides Link:
Cheng Jin David Wei Steven Low FAST TCP: design and experiments.
August 10, Circuit TCP (CTCP) Helali Bhuiyan
Web Server Benchmarking Using the Internet Protocol Traffic and Network Emulator Carey Williamson, Rob Simmonds, Martin Arlitt et al. University of Calgary.
Restricted Slow-Start for TCP William Allcock 1,2, Sanjay Hegde 3 and Rajkumar Kettimuthu 1,2 1 Argonne National Laboratory 2 The University of Chicago.
Ahmed El-Hassany CISC856: CISC 856 TCP/IP and Upper Layer Protocols Slides adopted from: Injong Rhee, Lisong Xu.
CUBIC : A New TCP-Friendly High-Speed TCP Variant Injong Rhee, Lisong Xu Member, IEEE v 0.2.
Enabling New Applications with Optical Circuit-Switched Networks Xuan Zheng April 27, 2004.
Cheng Jin David Wei Steven Low FAST TCP: Motivation, Architecture, Algorithms, Performance.
PFLDNet Argonne Feb 2004 R. Hughes-Jones Manchester 1 UDP Performance and PCI-X Activity of the Intel 10 Gigabit Ethernet Adapter on: HP rx2600 Dual Itanium.
TCP Westwood: Experiments over Large Pipes Cesar Marcondes Anders Persson Prof. M.Y. Sanadidi Prof. Mario Gerla NRL – Network Research Lab UCLA.
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji  Hemal V. Shah ¥ D. K. Panda 
Introduction 1 Lecture 14 Transport Layer (Congestion Control) slides are modified from J. Kurose & K. Ross University of Nevada – Reno Computer Science.
KEK Network Qi Fazhi KEK SW L2/L3 Switch for outside connections Central L2/L3 Switch A Netscreen Firewall Super Sinet Router 10GbE 2 x GbE IDS.
Masaki Hirabaru NICT The 3rd International HEP DataGrid Workshop August 26, 2004 Kyungpook National Univ., Daegu, Korea High Performance.
Experiences in Design and Implementation of a High Performance Transport Protocol Yunhong Gu, Xinwei Hong, and Robert L. Grossman National Center for Data.
Large File Transfer on 20,000 km - Between Korea and Switzerland Yusung Kim, Daewon Kim, Joonbok Lee, Kilnam Chon
High Performance User-Level Sockets over Gigabit Ethernet Pavan Balaji Ohio State University Piyush Shivam Ohio State University.
Best-Case WiBro Performance for a Single Flow 1 MICNET 2009 Shinae Woo †, Keon Jang †, Sangman Kim † Soohyun Cho *, Jaehwa Lee *, Youngseok Lee ‡, Sue.
Masaki Hirabaru ISIT and Genkai Genkai / Hyeonhae Workshop in Fukuoka Feb. 27, 2003 Performance of G / H Link.
Masaki Hirabaru Internet Architecture Group GL Meeting March 19, 2004 High Performance Data transfer on High Bandwidth-Delay Product Networks.
The Internet Hall of Fame Induction Memorial Party for the late Dr. Masaki Hirabaru Shigeki Goto JPNIC Wednesday 4 March,
Optimizing UDP-based Protocol Implementations Yunhong Gu and Robert L. Grossman Presenter: Michal Sabala National Center for Data Mining.
1 Masaki Hirabaru and Yasuhiro Koyama APEC-TEL APGrid Workshop September 6, 2005 e-VLBI: Science over High-Performance Networks.
E-VLBI over TransPAC Masaki HirabaruDavid LapsleyYasuhiro KoyamaAlan Whitney Communications Research Laboratory, Japan MIT Haystack Observatory, USA Communications.
Hirabaru, Koyama, and Kimura NICT APAN NOC Meeting January 20, 2005 e-VLBI updates.
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
Masaki Hirabaru CRL, Japan APAN Engineering Team Meeting APAN 2003 in Busan, Korea August 27, 2003 Common Performance Measurement Platform.
Measurement of Routing Switch-over Time with Redundant Link Masaki Hirabaru (NICT), Teruo Nakai (KDDI), Yoshitaka Hattori (KDDI), Motohiro Ishii (QIC),
Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.
APAN 10Gbps End-to-End Performance Measurement Masaki Hirabaru (NICT), Takatoshi Ikeda (KDDI/NICT), and Yasuichi Kitamura (NICT) July 19, 2006 Network.
A Measurement Based Memory Performance Evaluation of High Throughput Servers Garba Isa Yau Department of Computer Engineering King Fahd University of Petroleum.
E-VLBI and End-to-End Performance Masaki HirabaruYasuhiro KoyamaTetsuro Kondo NICT KoganeiNICT Kashima
High TCP performance over wide area networks Arlington, VA May 8, 2002 Sylvain Ravot CalTech HENP Working Group.
APII/APAN Measurement Framework: Advanced Network Observatory Masaki Hirabaru (NICT), Takatoshi Ikeda (KDDI Lab), Motohiro Ishii (QIC), and Yasuichi Kitamura.
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
Masaki Hirabaru Tsukuba WAN Symposium 2005 March 8, 2005 e-VLBI and End-to-End Performance over Global Research Internet.
HighSpeed TCP for High Bandwidth-Delay Product Networks Raj Kettimuthu.
Masaki Hirabaru Network Performance Measurement and Monitoring APAN Conference 2005 in Bangkok January 27, 2005 Advanced TCP Performance.
Data Transport Challenges for e-VLBI Julianne S.O. Sansa* * With Arpad Szomoru, Thijs van der Hulst & Mike Garret.
Masaki Hirabaru NICT Koganei 3rd e-VLBI Workshop October 6, 2004 Makuhari, Japan Performance Measurement on Large Bandwidth-Delay Product.
Masaki Hirabaru 情報処理学会四国支部講演会 December 17, 2004 インターネットでの長距離高性能データ転送.
ND The research group on Networks & Distributed systems.
TCP transfers over high latency/bandwidth networks Internet2 Member Meeting HENP working group session April 9-11, 2003, Arlington T. Kelly, University.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
CATNIP – Context Aware Transport/Network Internet Protocol Carey Williamson Qian Wu Department of Computer Science University of Calgary.
1 Masaki Hirabaru Network Architecture Group PL Meeting New Generation Network Research Center July 26, 2006 A Role of Network Architecture in e-VLBI.
Masaki Hirabaru NICT APAN JP NOC Meeting August 31, 2004 e-VLBI for SC2004 bwctl experiment with Internet2.
H. OhsakiITCom A control theoretical analysis of a window-based flow control mechanism for TCP connections with different propagation delays Hiroyuki.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
EDC Intenet2 AmericaView TAH Oct 28, 2002 AlaskaView Proof-of-Concept Test Tom Heinrichs, UAF/GINA/ION Grant Mah USGS/EDC Mike Rechtenbaugh USGS/EDC Jeff.
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
INDIANAUNIVERSITYINDIANAUNIVERSITY Status of FAST TCP and other TCP alternatives John Hicks TransPAC HPCC Engineer Indiana University APAN Meeting – Hawaii.
1 Masaki Hirabaru and Yasuhiro Koyama PFLDnet 2006 Febrary 2, 2006 International e-VLBI Experience.
FAST TCP Cheng Jin David Wei Steven Low netlab.CALTECH.edu GNEW, CERN, March 2004.
1 eVLBI Developments at Jodrell Bank Observatory Ralph Spencer, Richard Hughes- Jones, Simon Casey, Paul Burgess, The University of Manchester.
Masaki Hirabaru (NICT) and Jin Tanaka (KDDI) Impact of Bottleneck Queue on Long Distant TCP Transfer August 25, 2005 NOC-Network Engineering Session Advanced.
T3: TCP-based High-Performance and Congestion-aware Tunneling Protocol for Cloud Networking Satoshi Ogawa† Kazuki Yamazaki† Ryota Kawashima† Hiroshi Matsuo†
Performance Issues in 2010 PERT workshop 2010 Chris Welti
Transport Protocols over Circuits/VCs
e-VLBI Deployments with Research Internet
Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the
Transport protocols Outline: Xuan Zheng Univ. of Virginia
Review of Internet Protocols Transport Layer
Achieving reliable high performance in LFNs (long-fat networks)
Presentation transcript:

Masaki Hirabaru CRL, Japan ITRC MAI BoF Kochi November 6, 2003 広帯域・高遅延ネットワークでの TCP性能計測

Acknowledgements David Lapsley, Haystack Observatory, MIT Koyama Yasuhiro, Kashima Space Research Center, CRL Junichi Nakajima, Kashima Space Research Center, CRL Jouko Ritakari, Metsähovi Radio Observatory, HUT Katsushi Kobayashi, Information and Network Systems, CRL CRL RDNP / Tokyo XP / TransPAC / Abilene / Caltech

Backgrounds e-VLBI – geographically distributed observation, interconnecting radio antennas over the world Gigabit/real-time VLBI – multi-gigabit rate sampling

VLBI (Very Long Baseline Interferometry)

Motivations MIT Haystack – CRL Kashima e-VLBI Experiment on August 27, 2003 to measure UT1-UTC in 24 hours –41.54 GB CRL => MIT 107 Mbps (~50 mins) GB MIT => CRL 44.6 Mbps (~120 mins) –RTT ~220 ms, UDP throughput Mbps However TCP ~6-8 Mbps (per session, tuned) –BBFTP with 5 x 10 TCP sessions to gain performance HUT – CRL Kashima Gigabit VLBI Experiment -RTT ~325 ms, UDP throughput ~70 Mbps However TCP ~2 Mbps (as is), ~10 Mbps (tuned) -Netants (5 TCP sessions with ftp stream restart extension) They need high-speed / real-time / reliable / long-haul high-performance, huge data transfer.

Purpose Ensure >= 1 Gbps end-to-end performance in high bandwidth-delay product networks –to support for networked science applications –to help operations in finding a bottleneck –to evaluate advanced transport protocols (e.g. Tsunami, SABUL, HSTCP, FAST, XCP, ikob)

Contents Advanced TCP evaluation on TransPAC/Internet2 How to isolate host and implementation specific constraints TCP monitoring with web100 Advanced TCP evaluation in a laboratory Summary and future work

Kwangju Busan 2.5G Fukuoka Korea 2.5G SONET KOREN Taegu Daejon 10G 0.6G1Gx2 QGPOP Seoul XP Genkai XP Kitakyushu Tokyo XP Kashima 0.1G Fukuoka Japan 250km 1,000km 2.5G TransPAC 9,000km 4,000km Los Angeles Cicago New York MIT Haystack HUT 10G 1G APII/JGN Abilene 0.1G Helsinki 2.4G Stockholm 0.6G 2.4G GEANT Nordunet funet Koganei 1G 7,000km Indianapolis I2 Venue 1G 10G 100km server (general) server (e-VLBI) Abilene Observatory: servers at each NOC CMM: common measurement machines Network Diagram for TransPAC/I2 Measurement (Oct. 2003) 1G x2 sender receiver Mark5 Linux (RH 7.1) P3 1.3GHz Memory 256MB GbE SK-9843 PE1650 Linux (RH 9) Xeon 1.4GHz Memory 1GB GbE Intel Pro/1000 XT Iperf UDP ~900Mbps (no loss)

TransPAC/I2 #1:Reno (Win 64MB)

TransPAC/I2 #1:Reno (better case)

TransPAC/I2 #1:Reno (worse case)

TransPAC/I2 #1: Reno (10 mins)

TransPAC/I2 #2:High Speed (Win 64MB)

TransPAC/I2 #2:High Speed (better case)

TransPAC/I2 #2:High Speed (worse case)

TransPAC/I2 #2: High Speed (60 mins)

TransPAC/I2 #1: Reno (Win 12MB)

TransPAC/I2 #2: High Speed (Win 12MB)

*Testing FAST TCP over Abilene by Shalunov, Internet2 Meeting, Oct Path4: Pittsburgh to Atlanta, RTT=26.9ms

FASTLinux throughput loss queue STCPHSTCP 30min Room for mice ! HSTCP * TCP comparison (dummynet 800M, 120ms) from FAST TCP presented IETF Vienna July 2003

1 st Step: Tuning a Host with UDP Remove any bottlenecks on a host –CPU, Memory, Bus, OS (driver), … Dell PowerEdge 1650 (*not enough power) –Intel Xeon 1.4GHz x1(2), Memory 1GB –Intel Pro/1000 XT onboard PCI-X (133Mhz) Dell PowerEdge 2650 –Intel Xeon 2.8GHz x1(2), Memory 1GB –Intel Pro/1000 XT PCI-X (133Mhz) Iperf UDP throughput 957 Mbps –GbE wire rate: headers: UDP(20B)+IP(20B)+EthernetII(38B) –Linux (RedHat 9) with web100 –PE1650: TxIntDelay=0

Evaluating Advanced TCPs Reno (Linux TCP, web100 version) –Ack: w=w+1/w, Loss: w=w-1/2*w HighSpeed TCP (included in web100) –Ack: w=w+a(w)/w, Loss: w=w-1/b(w)*w FAST TCP (binary, provided from Caltech) –w=1/2*(w_old*baseRTT/avgRTT+α+w_current) Limited Slow-Start (included in web 100) Note: Differences in sender side only

Web100 ( A kernel patch for monitoring/modifying TCP metrics in Linux kernel Alpha 2.3 for Linux , released September 12, 2003

2 nd Step: Tuning a Host with TCP Maximum socket buffer size (TCP window size) –net.core.wmem_max net.core.rmem_max (64MB) –net.ipv4.tcp_wmem net.tcp4.tcp_rmem (64MB) Driver descriptor length –e1000: TxDescriptors=1024 RxDescriptors=256 (default) Interface queue length –txqueuelen=100 (default) –net.core.netdev_max_backlog=300 (default) Interface queue descriptor –fifo (default) MTU –mtu=1500 (IP MTU) Iperf TCP throughput 941 Mbps –GbE wire rate: headers: TCP(32B)+IP(20B)+EthernetII(38B) –Linux (RedHat 9) with web100 Web100 (incl. High Speed TCP) –net.ipv4.web100_no_metric_save=1 (do not store TCP metrics in the route cache) –net.ipv4.WAD_IFQ=1 (do not send a congestion signal on buffer full) –net.ipv4.web100_rbufmode=0 net.ipv4.web100_sbufmode=0 (disable auto tuning) –Net.ipv4.WAD_FloydAIMD=1 (HighSpeed TCP) –net.ipv4.web100_default_wscale=7 (default) FAST TCP from Caltech –Linux –net.ipv4.tcp_vegas_cong_avoid=1 –net.ipv4.tcp_vegas_pascing=1 net.ipv4.tcp_vegas_fast_converge=1 –net.ipv4.tcp_vegas_alpha=100 net.ipv4.tcp_vegas_beta=120 net.ipv4.tcp_vegas_gamma=50

Test in a laboratory (1) – no bottleneck Packet Sphere ReceiverSender L2SW (12GCF) Bandwidth 1Gbps Buffer 0KB Delay 88 ms Loss 0 GbE/SX GbE/T PE 2650PE 1650 #0: Reno => Reno –net.ipv4.WAD_IFQ=0 (congestion signal on buffer full) #1: Reno => Reno #2: High Speed TCP => Reno #3: FAST TCP => Reno 2*BDP = 11MB #0, 1, 2: Data obtained on sender #3: Data obtained on receiver

Laboratory #0: Reno (no bottleneck)

Laboratory #0: Reno (60 mins)

Laboratory #1: Reno (no bottleneck)

Laboratory #1: Reno (60 mins)

Laboratory #2: High Speed (no bottleneck)

Laboratory #2: High Speed (60 mins)

Laboratory #3: FAST (no bottleneck)

Laboratory #3: FAST (60 mins)

Test in a laboratory (2) – with bottleneck Packet Sphere ReceiverSender L2SW (12GCF) Bandwidth 800Mbps Buffer 256KB Delay 88 ms Loss 0 GbE/SX GbE/T PE 2650PE 1650 #4: Reno => Reno #5: High Speed TCP => Reno #6: FAST TCP => Reno 2*BDP = 11MB #4, 5: Data obtained on sender #6: Data obtained on receiver

Laboratory #4: Reno (800Mbps)

Laboratory #4: Reno (startup)

Laboratory #4: High Speed (800Mbps)

Laboratory #5: High Speed (startup)

Laboratory #5: High Speed ( )

Laboratory #5: High Speed (Win 12MB)

Laboratory #5: High Speed (Limited slow-start)

Laboratory #7: FAST (800Mbps)

Laboratory #7: FAST (startup)

Test in a laboratory (3) – with bottleneck Packet Sphere ReceiverSender L2SW (12GCF) Bandwidth 100Mbps Buffer 256KB Delay 88 ms Loss 0 GbE/SX GbE/T PE 2650PE 1650 #7: Reno => Reno #8: HighSpeed TCP => Reno #9: FAST TCP => Reno #10: Reno => Reno (100M shaping) 2*BDP = 11MB #7, 8, 10: Data obtained on sender #9: Data obtained on receiver

Laboratory #7: Reno (100Mbps)

Laboratory #7: Reno (10 mins)

Laboratory #8: High Speed (100Mbps)

Laboratory #8: High Speed (10 mins)

Laboratory #9: FAST (100Mbps)

Laboratory #9: FAST (10 mins)

Laboratory #10: Reno (100Mbps shaping)

Summary and Future work No significant differences under no bottlenecks –Packet loss With bottleneck, FAST would work better in case RTT increase can be detected –RED on the bottleneck –Queue length on the bottleneck –Signal from the bottleneck or from the receiver Appropriate window size to avoid loss –Autotuning? Mobile

Questions? See for VLBI