ORNL Net100 status July 31, 2002. UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory ORNL Net100 Focus Areas (first year) –TCP optimizations.

Slides:



Advertisements
Similar presentations
Helping TCP Work at Gbps Cheng Jin the FAST project at Caltech
Advertisements

Appropriateness of Transport Mechanisms in Data Grid Middleware Rajkumar Kettimuthu 1,3, Sanjay Hegde 1,2, William Allcock 1, John Bresnahan 1 1 Mathematics.
Doc.: IEEE /0604r1 Submission May 2014 Slide 1 Modeling and Evaluating Variable Bit rate Video Steaming for ax Date: Authors:
1 TCP Congestion Control. 2 TCP Segment Structure source port # dest port # 32 bits application data (variable length) sequence number acknowledgement.
BZUPAGES.COM 1 User Datagram Protocol - UDP RFC 768, Protocol 17 Provides unreliable, connectionless on top of IP Minimal overhead, high performance –No.
1 Web Server Performance in a WAN Environment Vincent W. Freeh Computer Science North Carolina State Vsevolod V. Panteleenko Computer Science & Engineering.
Chapter 3 Transport Layer slides are modified from J. Kurose & K. Ross CPE 400 / 600 Computer Communication Networks Lecture 12.
Transport Layer 3-1 Fast Retransmit r time-out period often relatively long: m long delay before resending lost packet r detect lost segments via duplicate.
High-performance bulk data transfers with TCP Matei Ripeanu University of Chicago.
1 Chapter 3 Transport Layer. 2 Chapter 3 outline 3.1 Transport-layer services 3.2 Multiplexing and demultiplexing 3.3 Connectionless transport: UDP 3.4.
Data Communication and Networks
Transport Level Protocol Performance Evaluation for Bulk Data Transfers Matei Ripeanu The University of Chicago Abstract:
Introduction 1 Lecture 14 Transport Layer (Congestion Control) slides are modified from J. Kurose & K. Ross University of Nevada – Reno Computer Science.
The Effects of Systemic Packets Loss on Aggregate TCP Flows Thomas J. Hacker May 8, 2002 Internet 2 Member Meeting.
Development of network-aware operating systems Tom Dunigan
Transport Layer 4 2: Transport Layer 4.
Transport Layer3-1 Chapter 3 outline r 3.1 Transport-layer services r 3.2 Multiplexing and demultiplexing r 3.3 Connectionless transport: UDP r 3.4 Principles.
Experiences in Design and Implementation of a High Performance Transport Protocol Yunhong Gu, Xinwei Hong, and Robert L. Grossman National Center for Data.
Maximizing End-to-End Network Performance Thomas Hacker University of Michigan October 26, 2001.
Implementing High Speed TCP (aka Sally Floyd’s) Yee-Ting Li & Gareth Fairey 1 st October 2002 DataTAG CERN (Kinda!)
UDT: UDP based Data Transfer Yunhong Gu & Robert Grossman Laboratory for Advanced Computing University of Illinois at Chicago.
1 Project Goals Project Elements Future Plans Scheduled Accomplishments Project Title: Net Developing Network-Aware Operating Systems PI: G. Huntoon,
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
Parallel TCP Bill Allcock Argonne National Laboratory.
HighSpeed TCP for High Bandwidth-Delay Product Networks Raj Kettimuthu.
1 Mao W07 Midterm Review EECS 489 Computer Networks Z. Morley Mao Monday Feb 19, 2007 Acknowledgement: Some.
Requirements for Simulation and Modeling Tools Sally Floyd NSF Workshop August 2005.
Rate Control Rate control tunes the packet sending rate. No more than one packet can be sent during each packet sending period. Additive Increase: Every.
NET100 Development of network-aware operating systems Tom Dunigan
UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory Net100 PIs: Wendy Huntoon/PSC, Tom Dunigan/ORNL, Brian Tierney/LBNL Impact and Connections.
Network-aware OS DOE/MICS Project Review August 18, 2003 Tom Dunigan Matt Mathis Brian Tierney
NET100 … as seen from ORNL Tom Dunigan November 8, 2001.
Iperf Quick Mode Ajay Tirumala & Les Cottrell. Sep 12, 2002 Iperf Quick Mode at LBL – Les Cottrell & Ajay Tirumala Iperf QUICK Mode Problem – Current.
NET100 Development of network-aware operating systems Tom Dunigan
Analysis of Buffer Size in Core Routers by Arthur Dick Supervisor Anirban Mahanti.
Network-aware OS DOE/MICS Project Final Review September 16, 2004 Tom Dunigan Matt Mathis Brian Tierney ORNL.
Lecture 9 – More TCP & Congestion Control
Self-generated Self-similar Traffic Péter Hága Péter Pollner Gábor Simon István Csabai Gábor Vattay.
Transport Layer 3-1 Chapter 3 Transport Layer Computer Networking: A Top Down Approach 6 th edition Jim Kurose, Keith Ross Addison-Wesley March
CS640: Introduction to Computer Networks Aditya Akella Lecture 15 TCP – III Reliability and Implementation Issues.
Web100/Net100 at Oak Ridge National Lab Tom Dunigan August 1, 2002.
CS640: Introduction to Computer Networks Aditya Akella Lecture 15 TCP – III Reliability and Implementation Issues.
Transport Layer3-1 Chapter 3 outline r 3.1 Transport-layer services r 3.2 Multiplexing and demultiplexing r 3.3 Connectionless transport: UDP r 3.4 Principles.
Transport Layer 3- Midterm score distribution. Transport Layer 3- TCP congestion control: additive increase, multiplicative decrease Approach: increase.
TCP OVER ADHOC NETWORK. TCP Basics TCP (Transmission Control Protocol) was designed to provide reliable end-to-end delivery of data over unreliable networks.
NET100 Development of network-aware operating systems Tom Dunigan
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory Net100: developing network-aware operating systems New (9/01) DOE-funded (Office of.
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
INDIANAUNIVERSITYINDIANAUNIVERSITY Status of FAST TCP and other TCP alternatives John Hicks TransPAC HPCC Engineer Indiana University APAN Meeting – Hawaii.
Fall 2004FSU CIS 5930 Internet Protocols1 Second phase of the project Please check some networking textbooks for details on TCP and OSPF.
Network-aware OS ESCC Miami February 5, 2003 Tom Dunigan Matt Mathis Brian Tierney
Network-aware OS DOE/MICS Project Review August 18, 2003 Tom Dunigan Matt Mathis Brian Tierney CSM lunch.
UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory Net100 year 1 leftovers (proposal): PSC –none ORNL –router access to SNMP data (besides.
Network-aware OS DOE/MICS ORNL site visit January 8, 2004 ORNL team: Tom Dunigan, Nagi Rao, Florence Fowler, Steven Carter Matt Mathis Brian.
Adding Explicit Congestion Notification (ECN) Capability to TCP's SYN/ACK Packets A. Kuzmanovic, A. Mondal, S. Floyd, and K.K. Ramakrishnan draft-ietf-tcpm-ecnsyn-02.txt.
Transport Layer session 1 TELE3118: Network Technologies Week 11: Transport Layer TCP Some slides have been taken from: r Computer Networking:
Network Protocols: Design and Analysis Polly Huang EE NTU
A TCP Tuning Daemon SC2002 November 19, 2002 Tom Dunigan Matt Mathis Brian Tierney
Chapter 3 outline 3.1 transport-layer services
Chapter 6 TCP Congestion Control
Protocols for Low Power
Transport Protocols over Circuits/VCs
Prepared by Les Cottrell & Hadrien Bullot, SLAC & EPFL, for the
Chapter 6 TCP Congestion Control
TCP flow and congestion control
Anant Mudambi, U. Virginia
Summer 2002 at SLAC Ajay Tirumala.
Using NetLogger and Web100 for TCP analysis
Presentation transcript:

ORNL Net100 status July 31, 2002

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory ORNL Net100 Focus Areas (first year) –TCP optimizations (WAD, AIMD, VMSS, ns, atou) –Network tool evaluation (iperf, webd, traced, java/web100) –Bulk transfer optimization, GridFTP (LBL/NERSC/ORNL) Today’s agenda –Activities since Denver meeting –Current activities / WAD status –Future work/needs

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory ORNL activities since Denver meeting Web100 tuning (manual) of HSI transfer (NERSC/ORNL), SC2001 NISTNet testbed SCTP study/report autotuning WAD (version 0) TCP tuning (AIMD, VMSS) with ns and atou Web100 event notification Web100 tools (ttcp100/iperf100, traced, tracer.py,webd,java bw tester) Evaluation of Linux 2.4 (tuning, caching, sendstall, del ACK, reorder) DRS testing (ESnet, NISTNet), integration with Web100 (2.4.16) TCP Vegas testing/porting (ns, Linux, atou) WAD tuning (AIMD, slow-start, buffers, NTAF data) (SC02 paper) GridFTP and parallel stream tuning Analysis of parallel-TCP and dynamic right-sizing ( ICN02 paper) Implementation of linear squares and nearest neighbor estimators with linear fuser Outreach: Net100 talks, web pages, atou tech report, SC02 Net100 paper,ICN02 paper, interactions with Claffy, Feng, Cottrell, Floyd, SCNM, Internet2 e2epi

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory Web100 tools Post-transfer statistics –Java bandwidth tester (53% have pkt loss) –ttcp100/iperf100 –Web100 daemon avoid modifying applications log designated paths/ports/variables Tracer daemon –collect Web100 variables at 0.1 second intervals –config file specifies source/port dest/port web100 variables (current/delta) –log to disk with timestamp and CID –plot/analyze flows/aggregates –C and python (LBL-based) # traced config file #local lport remote rport #v=value d=delta d PktsOut d PktsRetrans v CurrentCwnd v SampledRTT

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory WAD Version 1 –event-based (socket open/close) –config file with “tuning info” buffer sizes, AIMD, slow-start – periodic poll of NTAF (flakey) –static tuning -- value in config file –dynamic tuning use buffer sizes from NTAF divide buffer size among concurrent flows tune AIMD with Floyd table based on buffer size, or periodically during flow python WAD (based on LBL work) –polling WAD config file [bob] src_addr: src_port: 0 dst_addr: dst_port: 0 mode: 1 sndbuf: rcvbuf: wadai: 6 wadmd: 0.3 maxssth: 100 divide: 1 reorder: 9 delack: 0 floyd: 1

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory WAD tuning results (your mileage may vary …) Classic buffer tuning : ORNL to PSC, OC12, 80ms RTT network-challenged app. gets 10 Mbs same app., WAD/NTAF tuned buffer get 143 Mbs Is there a buffer size where you don’t get loss? … NOT Virtual MSS tune TCP’s additive increase (WAD_AI) add K segments per RTT during recovery k=6 like GigE jumboframe

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory WAD tuning Modified slow-start and VMSS ORNL to NERSC, OC12, 80 ms RTT often losses in slow start WAD tunes Floyd slowstart (WAD_MaxThresh) and AI (6) Floyd s-s little improvement under heavy congestion…. WAD tuned AIMD and slow start ORNL to CERN, OC?, 150ms RTT parallel streams AIMD (1/(2k),k) WAD tune single stream (0.125,4)

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory WAD tuning: Floyd AIMD Floyd AIMD adjust AIMD as function of cwnd (loss assumption) bigger cwnd: bigger increment, smaller reduction tested with ns and atou (continuous) WAD implementation pre-tune based on target buffer size (aggressive) continuous tuning (0.1 second) discrete rather than continuous add to Linux 2.4 (soon) How to select AIMD? Jumbo, parallel equivalent, Floyd, others ?

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory GridFTP tuning Can tuned single stream compete with parallel streams? Mostly not with “equivalence” tuning, but sometimes…. Testing on the real net is problematic. WAD can divide buffer among concurrent flows, tests inconclusive so far…. Is there a “congestion metric”? Per unit of time? Flow Mbs congestion re-xmits untuned tuned parallel untuned tuned parallel Data/plots from web100 tracer Buffers: 64K I/O, 4MB TCP (untuned 64K TCP: 8 mbs, 200s)

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory Net100 TCP tuning Reorder threshold seeing more out of order packets WAD tune a bigger reorder threshold Linux 2.4 does a good job already (caches reorder) LBL to ORNL (using our TCP-over-UDP) dup3 case had 289 retransmits, but all were unneeded! Delayed ACKs WAD could turn off delayed ACKs -- 2x improvement in recovery rate and slowstart linux 2.4 already turns off delayed ACKs for initial slow-start ns simulation: 500 mbs link, 80 ms RTT Packet loss early in slow start. Standard TCP with del ACK takes 10 minutes to recover!

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory In progress... WAD enhancement and testing –delayed ACK –Floyd AIMD (WAD and/or kernel) –tuning with NTAF data –distribution to other Net100 sites GridFTP tuning (ORNL/PSC/LBL) python WAD with netlink parallel stream tuning TCP optimization studies (AIMD, Vegas) Addition of postdoc

UT-BATTELLE U.S. Department of Energy Oak Ridge National Laboratory ORNL Yr 1 Milestones 1.Deploy Web100 at ORNL and NERSC nodes to develop Net100 expertise 2.Develop and demonstrate Web100-aware data transfer application for Probe/HPSS testing between NERSC and ORNL 3. Contribute to test and evaluation of existing end-to-end tools 4.Get access to ESnet ORNL and NERSC routers and investigate possible realtime feedback to application (e.g. using SNMP) 5. Explore transport optimizations for single TCP flows 6. Develop file transfer application/protocol to support out-of-order packet arrivals 7. Deploy a small emulator testbed to test transport protocol modifications and out-of-order resilient protocols/applications 8. Explore tuning the IBM/AIX 5.1 TCP stack and investigate extending it with Net100 mods 9. Test net100 tools on ESnet's OC48 testbed 10. Publish tools and tips on web page and in formal publications and presentations