Presentation is loading. Please wait.

Presentation is loading. Please wait.

1 Content-Aware Device Benchmarking Methodology (draft-hamilton-bmwg-ca-bench-meth-05) BMWG Meeting IETF-79 Beijing November 2010 Mike Hamilton

Similar presentations


Presentation on theme: "1 Content-Aware Device Benchmarking Methodology (draft-hamilton-bmwg-ca-bench-meth-05) BMWG Meeting IETF-79 Beijing November 2010 Mike Hamilton"— Presentation transcript:

1 1 Content-Aware Device Benchmarking Methodology (draft-hamilton-bmwg-ca-bench-meth-05) BMWG Meeting IETF-79 Beijing November 2010 Mike Hamilton mhamilton@breakingpoint.com BreakingPoint Systems

2 2 Agenda Charter Update Now What? Proposed Schedule Goals Reset Goals/Non-Goals Proposed Schedule Outstanding Questions Draft-05 highlights

3 3 Charter Update New classes of network devices that operate above the IP layer of the network stack require a new methodology to perform adequate benchmarking. Existing BMWG RFCs (RFC2647 and RFC3511) provides useful measurement and performance statistics, though they may not reflect the actual performance of the device when deployed in production networks. Operating within the limitations of the charter, namely blackbox characterization in laboratory environments, the BMWG will develop a methodology that more closely relates the performance of these devices to performance in an operational setting. In order to confirm or identify key performance characteristics, BMWG will solicit input from operations groups such as NANOG, RIP and APRICOT.

4 4 Now What? Define strategy for getting the work done Terminology Draft Methodology Draft Call for volunteers Would like to setup weekly meetings

5 5 Proposed Schedule December 2011 Terminology Draft for IESG Review Methodology Draft for IESG Review December 2010 -00 terminology draft January 2011 -00 methodology draft March – June 2011 Solicit WG feedback June 2011 Start WGLC

6 6 Goals Reset Create a series of benchmark tests to MOST accurately predict device performance under realistic conditions FOR A SPECIFIC SIMULATED NETWORK RFC 2544 Quotes Page 11, Section 18, “Multiple Frame Sizes” “The distribution MAY approximate the conditions on the network in which the DUT would be used.” “The authors do not have any idea how the results of such a test would be interpreted other than to directly compare multiple DUTs in some very specific simulated network”

7 7 Goals/Non-Goals Goals Repeatable Results Random/Repeatable Testbed to testbed Compare Multiple DUTs Non-Goals Replace RFC 2544/3511

8 8 Outstanding Questions Traffic Composition Basket? Methodology for selecting? Updatable? One network is different from another every day Security? Is it worth the trouble? Malformed Traffic What information is necessary to create repeatable ‘application layer’ flows? Type of flow, size of flow, etc Others?

9 9 Draft-05 Highlights and Reasons “Shell” Methodology More reproducible Added back ‘security’ Why shy away when application layer already difficult Maintain ‘fuzzing’ aspect Random but repeatable


Download ppt "1 Content-Aware Device Benchmarking Methodology (draft-hamilton-bmwg-ca-bench-meth-05) BMWG Meeting IETF-79 Beijing November 2010 Mike Hamilton"

Similar presentations


Ads by Google