Estimating TCP Latency Approximately with Passive Measurements Sriharsha Gangam, Jaideep Chandrashekar, Ítalo Cunha, Jim Kurose.

Slides:



Advertisements
Similar presentations
Florin Dinu T. S. Eugene Ng Rice University Inferring a Network Congestion Map with Traffic Overhead 0 zero.
Advertisements

New Directions in Traffic Measurement and Accounting Cristian Estan – UCSD George Varghese - UCSD Reviewed by Michela Becchi Discussion Leaders Andrew.
A Scalable and Reconfigurable Search Memory Substrate for High Throughput Packet Processing Sangyeun Cho and Rami Melhem Dept. of Computer Science University.
Data Streaming Algorithms for Accurate and Efficient Measurement of Traffic and Flow Matrices Qi Zhao*, Abhishek Kumar*, Jia Wang + and Jun (Jim) Xu* *College.
Multi-granular, multi-purpose and multi-Gb/s monitoring on off-the-shelf systems TELE9752 Group 3.
3/13/2012Data Streams: Lecture 161 CS 410/510 Data Streams Lecture 16: Data-Stream Sampling: Basic Techniques and Results Kristin Tufte, David Maier.
A Fast and Compact Method for Unveiling Significant Patterns in High-Speed Networks Tian Bu 1, Jin Cao 1, Aiyou Chen 1, Patrick P. C. Lee 2 Bell Labs,
Fine-Grained Latency and Loss Measurements in the Presence of Reordering Myungjin Lee, Sharon Goldberg, Ramana Rao Kompella, George Varghese.
Cloud Control with Distributed Rate Limiting Barath Raghavan, Kashi Vishwanath, Sriram Ramabhadran, Kenneth Yocum, and Alex C. Snoeren University of California,
Detecting DDoS Attacks on ISP Networks Ashwin Bharambe Carnegie Mellon University Joint work with: Aditya Akella, Mike Reiter and Srinivasan Seshan.
Sampling: Final and Initial Sample Size Determination
Flowlet Switching Srikanth Kandula Shan Sinha & Dina Katabi.
Fast, Memory-Efficient Traffic Estimation by Coincidence Counting Fang Hao 1, Murali Kodialam 1, T. V. Lakshman 1, Hui Zhang 2, 1 Bell Labs, Lucent Technologies.
Fundamentals of Computer Networks ECE 478/578 Lecture #21: TCP Window Mechanism Instructor: Loukas Lazos Dept of Electrical and Computer Engineering University.
Sampling and Flow Measurement Eric Purpus 5/18/04.
An Implementation and Experimental Study of the eXplicit Control Protocol (XCP) Yongguang Zhang and Tom Henderson INFOCOMM 2005 Presenter - Bob Kinicki.
1 Modeling and Emulation of Internet Paths Pramod Sanaga, Jonathon Duerig, Robert Ricci, Jay Lepreau University of Utah.
1 An Open Source Hardware Module for High-Speed Network Monitoring on NetFPGA NetFPGA European Developers Workshop 2010 Gianni Antichi, Stefano Giordano.
Hash Tables With Finite Buckets Are Less Resistant to Deletions Yossi Kanizo (Technion, Israel) Joint work with David Hay (Columbia U. and Hebrew U.) and.
Reverse Hashing for Sketch Based Change Detection in High Speed Networks Ashish Gupta Elliot Parsons with Robert Schweller, Theory Group Advisor: Yan Chen.
Towards a High-speed Router-based Anomaly/Intrusion Detection System (HRAID) Zhichun Li, Yan Gao, Yan Chen Northwestern.
User-level Internet Path Diagnosis R. Mahajan, N. Spring, D. Wetherall and T. Anderson.
Dream Slides Courtesy of Minlan Yu (USC) 1. Challenges in Flow-based Measurement 2 Controller Configure resources1Fetch statistics2(Re)Configure resources1.
BUFFALO: Bloom Filter Forwarding Architecture for Large Organizations Minlan Yu Princeton University Joint work with Alex Fabrikant,
Hash, Don’t Cache: Fast Packet Forwarding for Enterprise Edge Routers Minlan Yu Princeton University Joint work with Jennifer.
Not All Microseconds are Equal: Fine-Grained Per-Flow Measurements with Reference Latency Interpolation Myungjin Lee †, Nick Duffield‡, Ramana Rao Kompella†
NET-REPLAY: A NEW NETWORK PRIMITIVE Ashok Anand Aditya Akella University of Wisconsin, Madison.
Department of Computer Science and Engineering Applied Research Laboratory 1 A Hardware Based TCP/IP Processing Engine David V. Schuehler
SIGCOMM 2002 New Directions in Traffic Measurement and Accounting Focusing on the Elephants, Ignoring the Mice Cristian Estan and George Varghese University.
Scalable and Efficient Data Streaming Algorithms for Detecting Common Content in Internet Traffic Minho Sung Networking & Telecommunications Group College.
Compact Data Structures and Applications Gil Einziger and Roy Friedman Technion, Haifa.
Noise Can Help: Accurate and Efficient Per-flow Latency Measurement without Packet Probing and Time Stamping Dept. of Computer Science and Engineering.
Efficient Packet Classification with Digest Caches Francis Chang Wu-chang Feng Wu-chi Feng Kang Li.
Vladimír Smotlacha CESNET Full Packet Monitoring Sensors: Hardware and Software Challenges.
Large-Scale IP Traceback in High-Speed Internet : Practical Techniques and Theoretical Foundation Jun (Jim) Xu Networking & Telecommunications Group College.
Resource/Accuracy Tradeoffs in Software-Defined Measurement Masoud Moshref, Minlan Yu, Ramesh Govindan HotSDN’13.
A Formal Analysis of Conservative Update Based Approximate Counting Gil Einziger and Roy Freidman Technion, Haifa.
Jennifer Rexford Princeton University MW 11:00am-12:20pm Measurement COS 597E: Software Defined Networking.
Opportunistic Traffic Scheduling Over Multiple Network Path Coskun Cetinkaya and Edward Knightly.
1 Network Measurements and Sampling Nick Duffield, Carsten Lund, and Mikkel Thorup AT&T Labs-Research, Florham Park, NJ.
Efficient Cache Structures of IP Routers to Provide Policy-Based Services Graduate School of Engineering Osaka City University
© 2006 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Applying Syntactic Similarity Algorithms.
Access Delay Distribution Estimation in Networks Avideh Zakhor Joint work with: E. Haghani and M. Krishnan.
Computer Networking Lecture 18 – More TCP & Congestion Control.
The Bloom Paradox Ori Rottenstreich Joint work with Isaac Keslassy Technion, Israel.
Connect. Communicate. Collaborate Using Temporal Locality for a Better Design of Flow-oriented Applications Martin Žádník, CESNET TNC 2007, Lyngby.
An Efficient Gigabit Ethernet Switch Model for Large-Scale Simulation Dong (Kevin) Jin.
Emir Halepovic, Jeffrey Pang, Oliver Spatscheck AT&T Labs - Research
An Efficient Gigabit Ethernet Switch Model for Large-Scale Simulation Dong (Kevin) Jin.
Noise Can Help: Accurate and Efficient Per-flow Latency Measurement without Packet Probing and Time Stamping Michigan State University SIGMETRICS 14.
SCREAM: Sketch Resource Allocation for Software-defined Measurement Masoud Moshref, Minlan Yu, Ramesh Govindan, Amin Vahdat (CoNEXT’15)
© 2006 Andreas Haeberlen, MPI-SWS 1 Monarch: A Tool to Emulate Transport Protocol Flows over the Internet at Large Andreas Haeberlen MPI-SWS / Rice University.
BUFFALO: Bloom Filter Forwarding Architecture for Large Organizations Minlan Yu Princeton University Joint work with Alex Fabrikant,
SketchVisor: Robust Network Measurement for Software Packet Processing
A Resource-minimalist Flow Size Histogram Estimator
The Variable-Increment Counting Bloom Filter
Inferring Queue Sizes in Access Networks by Active Measurement
FAST TCP : From Theory to Experiments
Advanced Computer Networks
Hash Functions for Network Applications (II)
pathChirp Efficient Available Bandwidth Estimation
Lecture 1: Bloom Filters
Author: Yi Lu, Balaji Prabhakar Publisher: INFOCOM’09
A flow aware packet sampling mechanism for high speed links
Lu Tang , Qun Huang, Patrick P. C. Lee
Toward Self-Driving Networks
pathChirp Efficient Available Bandwidth Estimation
Toward Self-Driving Networks
Author: Ramana Rao Kompella, Kirill Levchenko, Alex C
Presentation transcript:

Estimating TCP Latency Approximately with Passive Measurements Sriharsha Gangam, Jaideep Chandrashekar, Ítalo Cunha, Jim Kurose

Motivation: Network Troubleshooting Home network or access link? o WiFi signal or access link? ISP, enterprise and content provider networks o Within the network or outside? 2 ? ? ? Latency and network performance problems? Where? Why?

Passive Measurement in the Middle Decompose path latency of TCP flows o Latency: Important interactive web-applications o Other approaches: active probing, NetFlow Challenges o Constrained devices and high data rates 3 Challenge: Constrained Resources Passive measurement device (gateway, router) A B

TCP Latency 4 Existing methods: Emulate TCP state for each flow Accurate estimates but expensive SourceDestination Time T S1 T M1 T M2 T S2 T D1 T D2 RTT = T M2 – T M1 S S 0, 99 A A 100

TCP Latency 5 Scaling challenges o Large packet and flow arrivals o Constrained devices (limited memory/CPU) ALE – Approximate Latency Estimator SourceDestination Time Large TCP state Expensive lookup

Design Space 6 Accuracy Overhead Ideal solution ALE Goals: Configurable tradeoff Existing solutions e.g., tcptrace Expensive for constrained devices

ALE Overview 7 Eliminate packet timestamps o Group into time intervals with granularity (w) Store expected ACK o SEG: 0-99, ACK: 100 Use probabilistic set membership data structure o Counting Bloom Filters SourceDestination Time ΔT = w S S 0, 99 A A 100

Sliding window of buckets (time intervals) o Buckets contain a counting bloom filter (CBF) ALE Overview 8 CBF T ALE W w ww CBF S S 0, 99 A A 100 S S 0, 99 hash(FlowId, 99+1) i i CBF A A 100 i i RTT = Δ + w + w/2 hash(inv(FlowId), 100) Current Bucket Lookup i i i i i i Match w w/2Δ Missed Latencies Max Error: w/2

Controlling Error with ALE Parameters 9 Decrease w: Higher Accuracy Increase W: Higher Coverage ALE-U (uniform) W w Number of buckets = W/w

ALE-Exponential (ALE-E) Process large and small latency flows simultaneously Absolute error is proportional to the latency Larger buckets shift slowly 10 A A w 2w 4w B B C C D D E E BA DC F F DCBA G G Counting Bloom Filter (CBF) Merge

Error Sources Bloom filters are probabilistic structures o False positives and negatives Artifacts from TCP Retransmitted packets and Reordered packets o Excess (to tcptrace) erroneous RTT samples ACK numbers not on SEQ boundaries, Cumulative ACKs 11

Evaluation Methodology 2 x Tier 1 backbone link traces (60s each) from CAIDA o Flows: 2,423,461, Packets: 54,089,453 o Bi-directional Flows: 93,791, Packets: 5,060,357 Ground truth/baseline comparison: tcptrace o Emulates TCP state machine Latencies ALE and tcptrace o Packet and flow level Overhead of ALE and tcptrace o Memory and compute 12

Latency Distribution ms120 ms 300 ms 500 ms Majority of the latencies

Accuracy: RTT Samples Range 0-60 ms Box plot of RTT differences More memory => Closer to tcptrace 14 Exact match with tcptrace Increasing overhead (smaller ‘w’)

Accuracy: RTT Samples 15 Exact match with tcptrace Increasing overhead ALE-E(12) outperforms ALE-U(24) for small latencies

Accuracy: Flow Statistics 16 Many applications use aggregated flow statics Small w (more buckets) o Median flow latencies approach tcptrace Exact match with tcptrace CDF

Accuracy: Flow Statistics 17 Small w: Standard deviation of flow latencies approach tcptrace Exact match with tcptrace

Compute and Memory Overhead Sample flows uniformly at random at rates 0.1, 0.2, o 5 pcap sub-traces per rate o Higher sampling rate (data rate) ⇒ More state for tcptrace GNU Linux taskstats API o Compute time and memory usage 18

Compute Time 19 ALE scales with increasing data rates

Memory Overhead TCPTRACEALE-U(96) 20 RSS memory: ≈64 MB (0.1) to ≈460 MB (0.7) VSS memory: ≈74 MB (0.1) to ≈468 MB (0.7) RSS memory: 2.0 MB for all Sampling rates VSS memory: 9.8 MB for all sampling rates ALE uses a fixed size data structure

Conclusions Current TCP measurements emulate state o Expensive: high data rates, constrained devices ALE provides low overhead data structure o Sliding window of CBF buckets Improvements over compute and memory with tcptrace sacrificing small accuracy o Tunable parameters Simple hash functions and counters o Hardware Implementation 21

Thank You 22

ALE Compute and Memory Bounds Insertion: O(h) time o CBF with h hash functions Matching ACK number: O(hn) time o ‘n’ buckets Shifting buckets: O(1) time o Linked list of buckets ALE-E: O(C) time to merge CBFs o ‘C’ counters in the CBF Memory usage: n × C × d bits o ‘d‘ bit CBF counters 23

ALE Error Bounds ALE-U: Average case error w/4 ALE-U: Worst case error w/2 ALE-E: Average case error (3w/16)2 i ALE-E: Worst case error (2 i−1 w) o ACK has a match in bucket i 24

ALE Parameters ‘w’ on accuracy requirements ‘W’ estimate of the maximum latencies expected CBFs with h = 4 hash functions C on traffic rate, w, false positive rate and h E.g., For w = 20 ms, h = 4, and m = R × w, the constraint for optimal h (h = (C/m) ln 2) yields C = counters 25

Compute Time 26 ALE scales with increasing data rates 10 min trace

Accuracy: RTT Samples 27

Memory Overhead 28

Eliminate packet timestamps o Time quantization: Use fixed number of intervals Store expected acknowledgement number ALE Overview 29 S S 0, 99 S S 100, 199 S S 200, 299 S S 300, 399 S S 400, 499 S S 500, 599 S S 600, 699 S S 700, 799 S S 800, 899 T1T1 T2T2 T3T3 S S 100 S S 200 S S 300 S S 400 S S 500 S S 600 S S 700 S S 800 S S 900 T1T1 T 2 = T 1 + w T 3 = T 1 + 2w

Motivation: Network Troubleshooting 30 Latency Sensitive Applications Latency and network performance problems? Where? Why?