CPSC 641: Network Traffic Self-Similarity

Slides:



Advertisements
Similar presentations
Autocorrelation Functions and ARIMA Modelling
Advertisements

Doc.: IEEE /1216r1 Submission November 2009 BroadcomSlide 1 Internet Traffic Modeling Date: Authors: NameAffiliationsAddressPhone .
Properties of Least Squares Regression Coefficients
Copyright © 2005 Department of Computer Science CPSC 641 Winter Self-Similar Network Traffic The original paper on network traffic self-similarity.
COMM 472: Quantitative Analysis of Financial Decisions
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Network and Service Assurance Laboratory Analysis of self-similar Traffic Using Multiplexer & Demultiplexer Loaded with Heterogeneous ON/OFF Sources Huai.
Time Series Building 1. Model Identification
2014 Examples of Traffic. Video Video Traffic (High Definition) –30 frames per second –Frame format: 1920x1080 pixels –24 bits per pixel  Required rate:
Finding Self-similarity in People Opportunistic Networks Ling-Jyh Chen, Yung-Chih Chen, Paruvelli Sreedevi, Kuan-Ta Chen Chen-Hung Yu, Hao Chu.
2  Something “feels the same” regardless of scale 4 What is that???
1 Self-Similar Ethernet LAN Traffic Carey Williamson University of Calgary.
CMPT 855Module Network Traffic Self-Similarity Carey Williamson Department of Computer Science University of Saskatchewan.
On the Self-Similar Nature of Ethernet Traffic - Leland, et. Al Presented by Sumitra Ganesh.
Stochastic Processes Dr. Talal Skaik Chapter 10 1 Probability and Stochastic Processes A friendly introduction for electrical and computer engineers Electrical.
October 14, 2002MASCOTS Workload Characterization in Web Caching Hierarchies Guangwei Bai Carey Williamson Department of Computer Science University.
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
On the Constancy of Internet Path Properties Yin Zhang, Nick Duffield AT&T Labs Vern Paxson, Scott Shenker ACIRI Internet Measurement Workshop 2001 Presented.
無線區域網路中自我相似交通流量之 成因與效能評估 The origin and performance impact of self- similar traffic for wireless local area networks 報 告 者:林 文 祺 指導教授:柯 開 維 博士.
Probability By Zhichun Li.
A Nonstationary Poisson View of Internet Traffic T. Karagiannis, M. Molle, M. Faloutsos University of California, Riverside A. Broido University of California,
Self-Similarity in Network Traffic Kevin Henkener 5/29/2002.
1 Interesting Links
Copyright © 2005 Department of Computer Science CPSC 641 Winter Network Traffic Measurement A focus of networking research for 20+ years Collect.
CSE 561 – Traffic Models David Wetherall Spring 2000.
Self-Similar through High-Variability: Statistical Analysis of Ethernet LAN Traffic at the Source Level Walter Willinger, Murad S. Taqqu, Robert Sherman,
CS 6401 Network Traffic Characteristics Outline Motivation Self-similarity Ethernet traffic WAN traffic Web traffic.
So are how the computer determines the size of the intercept and the slope respectively in an OLS regression The OLS equations give a nice, clear intuitive.
Self-Similar Traffic COMP5416 Advanced Network Technologies.
Self-Similarity of Network Traffic Presented by Wei Lu Supervised by Niclas Meier 05/
1 Chapters 9 Self-SimilarTraffic. Chapter 9 – Self-Similar Traffic 2 Introduction- Motivation Validity of the queuing models we have studied depends on.
References for M/G/1 Input Process
Network Traffic Modeling Punit Shah CSE581 Internet Technologies OGI, OHSU 2002, March 6.
Dr. Nawaporn Wisitpongphan. What do you need in order to conduct a PhD Research? A lot of reading in order to Find a PhD thesis topic Find out what other.
1 Statistical Distribution Fitting Dr. Jason Merrick.
1 Self Similar Traffic. 2 Self Similarity The idea is that something looks the same when viewed from different degrees of “magnification” or different.
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
K. Ensor, STAT Spring 2004 Memory characterization of a process How would the ACF behave for a process with no memory? What is a short memory series?
CY1B2 Statistics1 (ii) Poisson distribution The Poisson distribution resembles the binomial distribution if the probability of an accident is very small.
Correlation Forensic Statistics CIS205. Introduction Chi-squared shows the strength of relationship between variables when the data is of count form However,
1 Interesting Links. On the Self-Similar Nature of Ethernet Traffic Will E. Leland, Walter Willinger and Daniel V. Wilson BELLCORE Murad S. Taqqu BU Analysis.
Modeling and Simulation CS 313
CLT and Levy Stable Processes John Rundle Econophysics PHYS 250
Inference about the slope parameter and correlation
Stochastic Process - Introduction
MATH-138 Elementary Statistics
Regression Analysis AGEC 784.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Queuing Theory and Traffic Analysis
Linear Regression.
Modeling and Simulation CS 313
Ch9 Random Function Models (II)
Internet Traffic Modeling
CPSC 531: System Modeling and Simulation
Interesting Links.
Minimal Envelopes.
CPSC 641: Network Measurement
Statistical Methods Carey Williamson Department of Computer Science
Notices of the AMS, September 1998
CPSC 531: System Modeling and Simulation
CPSC 531: System Modeling and Simulation
Data Presentation Carey Williamson Department of Computer Science
Self-similar Distributions
Presented by Chun Zhang 2/14/2003
Continuous Growth and the Natural Exponential Function
Carey Williamson Department of Computer Science University of Calgary
Carey Williamson Department of Computer Science University of Calgary
Chapter 6 Random Processes
CPSC 641: Network Measurement
CH2 Time series.
Presentation transcript:

CPSC 641: Network Traffic Self-Similarity Carey Williamson Department of Computer Science University of Calgary

Introduction A classic network traffic measurement study has shown that aggregate Ethernet LAN traffic is self-similar [Leland et al 1993] A statistical property that is very different from the traditional Poisson-based models This presentation: definition of network traffic self-similarity, Bellcore Ethernet LAN data, implications of self-similarity

Measurement Methodology Collected lengthy traces of Ethernet LAN traffic on Ethernet LAN(s) at Bellcore High resolution time stamps Analyzed statistical properties of the resulting time series data Each observation represents the number of packets (or bytes) observed per time interval (e.g., 10 4 8 12 7 2 0 5 17 9 8 8 2...)

Self-Similarity: The Intuition If you plot the number of packets observed per time interval as a function of time, then the plot looks ‘‘similar’’ regardless of what interval size you choose E.g., 10 msec, 100 msec, 1 sec, 10 sec,... Same applies if you plot number of bytes observed per interval of time

Self-Similarity: The Intuition In other words, self-similarity implies a ‘‘fractal-like’’ behaviour: no matter what time scale you use to examine the data, you see similar patterns Implications: Burstiness exists across many time scales No natural length of a burst Traffic does not necessarilty get ‘‘smoother” when you aggregate it (unlike Poisson traffic)

Self-Similarity: The Mathematics Self-similarity is a rigourous statistical property (i.e., a lot more to it than just the pretty ‘‘fractal-like’’ pictures) Assumes you have time series data with finite mean and variance (i.e., covariance stationary stochastic process) Must be a very long time series (infinite is best!) Can test for presence of self-similarity

Self-Similarity: The Mathematics Self-similarity manifests itself in several equivalent fashions: Slowly decaying variance Long range dependence Non-degenerate autocorrelations Hurst effect

Slowly Decaying Variance The variance of the sample decreases more slowly than the reciprocal of the sample size For most processes, the variance of a sample diminishes quite rapidly as the sample size is increased, and stabilizes soon For self-similar processes, the variance decreases very slowly, even when the sample size grows quite large

Variance-Time Plot The ‘‘variance-time plot” is one method to test for the slowly decaying variance property Plots the variance of the sample versus the sample size, on a log-log plot For most processes, the result is a straight line with slope -1 For self-similar, the line is much flatter

Variance-Time Plot Variance m

Variance m Variance-Time Plot Variance of sample 100.0 10.0 Variance of sample on a logarithmic scale Variance 0.01 0.001 0.0001 m

Variance m Variance-Time Plot Sample size m on a logarithmic scale 4 5 6 7 1 10 100 10 10 10 10

Variance-Time Plot Variance m

Variance-Time Plot Variance m

Variance-Time Plot Slope = -1 for most processes Variance m

Variance-Time Plot Variance m

Variance m Variance-Time Plot Slope flatter than -1 for self-similar process m

Long Range Dependence Correlation is a statistical measure of the relationship, if any, between two random variables Positive correlation: both behave similarly Negative correlation: behave in opposite fashion No correlation: behaviour of one is statistically unrelated to behaviour of other

Long Range Dependence (Cont’d) Autocorrelation is a statistical measure of the relationship, if any, between a random variable and itself, at different time lags Positive correlation: big observation usually followed by another big, or small by small Negative correlation: big observation usually followed by small, or small by big No correlation: observations unrelated

Long Range Dependence (Cont’d) Autocorrelation coefficient can range between +1 (very high positive correlation) and -1 (very high negative correlation) Zero means no correlation Autocorrelation function shows the value of the autocorrelation coefficient for different time lags k

Autocorrelation Function +1 Autocorrelation Coefficient -1 lag k 100

Autocorrelation Function +1 Maximum possible positive correlation Autocorrelation Coefficient -1 lag k 100

Autocorrelation Function +1 Autocorrelation Coefficient Maximum possible negative correlation -1 lag k 100

Autocorrelation Function +1 No observed correlation at all Autocorrelation Coefficient -1 lag k 100

Autocorrelation Function +1 Autocorrelation Coefficient -1 lag k 100

Autocorrelation Function +1 Significant positive correlation at short lags Autocorrelation Coefficient -1 lag k 100

Autocorrelation Function +1 Autocorrelation Coefficient No statistically significant correlation beyond this lag -1 lag k 100

Long Range Dependence (Cont’d) For most processes (e.g., Poisson, or compound Poisson), the autocorrelation function drops to zero very quickly (usually immediately, or exponentially fast) For self-similar processes, the autocorrelation function drops very slowly (i.e., hyperbolically) toward zero, but may never reach zero Non-summable autocorrelation function

Autocorrelation Function +1 Typical long-range dependent process Autocorrelation Coefficient Typical short-range dependent process -1 lag k 100

Non-Degenerate Autocorrelations For self-similar processes, the autocorrelation function for the aggregated process is indistinguishable from that of the original process If autocorrelation coefficients match for all lags k, then called exactly self-similar If autocorrelation coefficients match only for large lags k, then called asymptotically self-similar

Autocorrelation Function +1 Original self-similar process Autocorrelation Coefficient Aggregated self-similar process -1 lag k 100

Aggregation Aggregation of a time series X(t) means smoothing the time series by averaging the observations over non-overlapping blocks of size m to get a new time series X (t) m

Aggregation: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1... Then the aggregated series for m = 2 is:

Aggregation: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1... Then the aggregated series for m = 2 is: 4.5 8.0 2.5 5.0 6.0 7.5 7.0 4.0 4.5 5.0...

Aggregation: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1... Then the aggregated time series for m = 5 is:

Aggregation: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1... Then the aggregated time series for m = 5 is: 6.0 4.4 6.4 4.8 ...

Autocorrelation Function +1 Original self-similar process Autocorrelation Coefficient Aggregated self-similar process -1 lag k 100

R(n) = max(0, W , ... W ) - min(0, W , ... W ) The Hurst Effect For almost all naturally occurring time series, the rescaled adjusted range statistic (also called the R/S statistic) for sample size n obeys the relationship E[R(n)/S(n)] = c n where: R(n) = max(0, W , ... W ) - min(0, W , ... W ) S(n) is the sample standard deviation, and W = X - k X for k = 1, 2, ... n H 1 n 1 n k k i n i =1

The Hurst Effect (Cont’d) For models with only short range dependence, H is almost always 0.5 For self-similar processes, 0.5 < H < 1.0 This discrepancy is called the Hurst Effect, and H is called the Hurst parameter Single parameter to characterize self-similar process!

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 There are 20 data points in this example

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 There are 20 data points in this example For R/S analysis with n = 1, you get 20 samples, each of size 1: (boring base case) Block 1: X = 2, W = 0, R(n) = 0, S(n) = 0 Block 2: X = 7, W = 0, R(n) = 0, S(n) = 0 Block 3: X = 4, W = 0, R(n) = 0, S(n) = 0 1 1 1

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 For R/S analysis with n = 2, you get 10 samples, each of size 2:

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 For R/S analysis with n = 2, you get 10 samples, each of size 2: Block 1: X = 4.5, W = -2.5, W = 0, R(n) = 0 - (-2.5) = 2.5, S(n) = 2.5, R(n)/S(n) = 1.0 n 1 2

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 For R/S analysis with n = 2, you get 10 samples, each of size 2: Block 2: X = 8.0, W = -4.0, W = 0, R(n) = 0 - (-4.0) = 4.0, S(n) = 4.0, R(n)/S(n) = 1.0 n 1 2

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 For R/S analysis with n = 5, you get 4 samples, each of size 5:

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 For R/S analysis with n = 5, you get 4 samples, each of size 4: Block 1: X = 6.0, W = -4.0, W = -3.0, W = -5.0 , W = 1.0 , W = 0, S(n) = 3.41, R(n) = 1.0 - (-5.0) = 6.0, R(n)/S(n) = 1.76 n 1 2 3 4 5

R/S Statistic: An Example Suppose the original time series X(t) contains the following (made up) values: 2 7 4 12 5 0 8 2 8 4 6 9 11 3 3 5 7 2 9 1 For R/S analysis with n = 5, you get 4 samples, each of size 4: Block 2: X = 4.4, W = -4.4, W = -0.8, W = -3.2 , W = 0.4 , W = 0, S(n) = 3.2, R(n) = 0.4 - (-4.4) = 4.8, R(n)/S(n) = 1.5 n 1 2 3 4 5

R/S Plot Another way of testing for self-similarity, and estimating the Hurst parameter Plot the R/S statistic for different values of n, with a log scale on each axis If time series is self-similar, the resulting plot will have a straight line shape with a slope H that is greater than 0.5 Called an R/S plot, or R/S pox diagram

R/S Pox Diagram R/S Statistic Block Size n

R/S Statistic Block Size n R/S Pox Diagram R/S statistic R(n)/S(n) on a logarithmic scale R/S Statistic Block Size n

R/S Statistic Block Size n R/S Pox Diagram Sample size n on a logarithmic scale Block Size n

Slope 1.0 R/S Statistic Slope 0.5 Block Size n R/S Pox Diagram Slope H (Hurst parameter) Slope 1.0 R/S Statistic Slope 0.5 Block Size n

Summary Self-similarity is an important mathematical property that has been identified as present in network traffic measurements Important property: burstiness across many time scales, traffic does not aggregate well There exist several mathematical methods to test for the presence of self-similarity, and to estimate the Hurst parameter H There exist models for self-similar traffic