Presentation is loading. Please wait.

Presentation is loading. Please wait.

Internet2 E2E piPEs Update

Similar presentations


Presentation on theme: "Internet2 E2E piPEs Update"— Presentation transcript:

1 Internet2 E2E piPEs Update
Eric L. Boyd 25 February 2019

2 Overview E2E piPEs Overview LA  CERN Measurement Demo piPEs Status
What is piPEs? Goals E2E piPEs Measurement Infrastructure LA  CERN Measurement Demo piPEs Status Where are we? What’s available? What’s next? What can you do? 2/25/2019

3 Internet2 E2E piPEs Project: End-to-End Performance Initiative Performance Environment System (E2E piPEs) Approach: Collaborative project combining the best work of many organizations, including DANTE/GEANT, Daresbury, EGEE, GGF NMWG, NLANR/DAST, UCL, Georgia Tech, etc. 2/25/2019

4 Internet2 E2E piPEs Goals
Enable end-users & network operators to: determine E2E performance capabilities locate E2E problems contact the right person to get an E2E problem resolved. Enable remote initiation of partial path performance tests Make partial path performance data publicly available Interoperable with other performance measurement frameworks 2/25/2019

5 Measurement Infrastructure Components
For any given “partial path” segment (dotted black line) … We might run regularly scheduled tests and/or on-demand tests. This means, we require the ability to make a test request, have test results stored in a database, request those test resutls, and retrieve the test results. 2/25/2019

6 Sample piPEs Deployment
Deployment is an inside-out approach. Start with regularly scheduled tests inside, make sure it plays well with regularly scheduled tests outside. Hope that projects working on the end nodes will meet us in the middle. 2/25/2019

7 Measurement Software Components
Boxes in black working and deployed (either released or in prototype form). Boxes in red under development. Theses are the software components that make up the piPEs measurement framework. Some are released (BWCTL, OWAMP), some are in “prototype format” (Database, Traceroute, PMP, PMC, web service, network monitoring), and some are under development (“Detective Applet”, Discovery module, Analysis module, MDI, NDT). The Measurement Domain Interface (MDI) is a web services interface that speaks the GGF NMWG Request/Report schema and handles authentication and authorization. It is being designed to be interoperable with other measurement frameworks (current and future). The Network Diagnostic Tool (NDT) is an existing tool that the original author is integrating into piPEs. It is designed to detect common problems in the first mile (the common case for most network “issues”). 2/25/2019

8 piPEs Deployment 1) Abilene Backbone Deployment (Complete)
Hawaii 2) Hawaii Campus Deployment (Complete) OSU NC State Europe UCSD 3) Campus, European, APAN Deployment (In Progress) Abilene Backbone deployment in Green. Campus deployment in Solid Red. In progress roll-out to Campuses and European partners in Dotted Red. “Europe” in this case means a handful of nodes scattered across the GEANT backbone and some nodes at CERN. We are trying to work very closely with DANTE/GEANT and EGEE to ensure interoperability (at a minimum) if not a shared code base across the US and Europe. This is important as many scientists in the US routinely ship data across the Atlantic and thus their E2E concerns are international in nature. 2/25/2019

9 Overview E2E piPEs Overview LA  CERN Measurement Demo piPEs Status
What is piPEs? Goals E2E piPEs Measurement Infrastructure LA  CERN Measurement Demo piPEs Status Where are we? What’s available? What’s next? What can you do? 2/25/2019

10 American/European Collaboration Goals
Awareness of ongoing Measurement Framework Efforts / Sharing of Ideas (Good / Not Sufficient) Interoperable Measurement Frameworks (Minimum) Common means of data extraction Partial path analysis possible along transatlantic paths Open Source Shared Development (Possibility, In Whole or In Part) End-to-end partial path analysis for transatlantic research communities VLBI: Onsala, Sweden  Haystack, Mass. HENP: CERN, Switzerland  Caltech, Calif. 2/25/2019

11 American/European Demonstration Goals
Demonstrate ability to do partial path analysis between “Caltech” (Los Angeles Abilene router) and CERN. Demonstrate ability to do partial path analysis involving nodes in the GEANT network. Compare and contrast measurement of a “lightpath” versus a normal IP path. Demonstrate interoperability of piPEs and analysis tools such as Advisor and MonALISA 2/25/2019

12 Demonstration Details
Path 1: Default route between LA and CERN is across Abilene to Chicago, then across Datatag circuit to CERN Path 2: Announced addresses so that route between LA and CERN traverses GEANT via London node Path 3: “Lightpath” (discussed earlier by Rick Summerhill) Each measurement “node” consists of a BWCTL box and an OWAMP box “next to” the router. 2/25/2019

13 All Roads Lead to Geneva
2/25/2019

14 Results BWCTL: OWAMP: MONALISA NLANR Advisor 2/25/2019

15 Insights (1) Even with shared source and a single team of developer-installers, inter-administrative domain coordination is difficult. Struggled with basics of multiple paths. IP addresses, host configuration, software (support source addresses, etc.) Struggled with cross-domain administrative coordination issues. AA (accounts), routes, port filters, MTUs, etc. Struggled with performance tuning measurement nodes. host tuning, asymmetric routing, MTUs We had log-in access … still struggled with IP addresses, accounts, port filters, host tuning, host configuration (using the proper paths), software. Port filters, MTUs. 2/25/2019

16 Insights (2) Connectivity takes a large amount of coordination and effort; performance takes even more of the same. Current measurement approaches have limited visibility into “lightpaths.” Having hosts participate in the measurement is one possible solution. 2/25/2019

17 Insights (3) Consider interaction with security; lack of end-to-end transparency is problematic. Security filters are set up based on expected traffic patterns Measurement nodes create new traffic Lightpaths bypass expected ingress points 2/25/2019

18 Overview E2E piPEs Overview LA  CERN Measurement Demo piPEs Status
What is piPEs? Goals E2E piPEs Measurement Infrastructure LA  CERN Measurement Demo piPEs Status Where are we? What’s available? What’s next? What can you do? 2/25/2019

19 piPEs Status Tools Measurement Framework Outreach
BWCTL (released) Traceroute (ongoing) OWAMP (released) NDT/piPEs Integration (ongoing) Measurement Framework Abilene Measurement Infrastructure (deployed) Measurement Infrastructure “Glue” (prototype, ongoing) Edge-to-Middle Testing (deployed) Outreach Application Communities (HENP, VLBI, ongoing) Establishing Measurement Domains (ITECs, ongoing) Collaboration Partners (DANTE, ESnet, APAN, ongoing) Standards Integration (GGF NMWG [Advisor, MonALISA], IETF, ongoing) Measurement Framework Federation 2/25/2019

20 BWCTL (Jeff Boote) http://e2epi.internet2.edu/bwctl
2/25/2019

21 OWAMP (Jeff Boote) http://e2epi.internet2.edu/owamp
2/25/2019

22 NDT (Rich Carlson) Network Diagnostic Tester
Developed at Argonne National Lab Ongoing integration into piPEs framework Redirects from well-known host to “nearest” measurement node Detects common performance problems in the “first mile” (edge to campus DMZ) In deployment on Abilene: 2/25/2019

23 Abilene Measurement Domain
Part of the Abilene Observatory: Regularly scheduled OWAMP (1-way latency) and BWCTL (Iperf wrapper) Tests Web pages displaying: Latest results “Weathermap” Worst 10 Performing Links Data available via web service: The E2E team is building the piPEs measurement framework. Internet2 has deployed an instance of that framework, the Abilene Measurement Domain (AMD). AMD is part of the Abilene Observatory. Currently, the AMD consists of regularly scheduled OWAMP and BWCTL tests, plus the ability of a user “on the edge” to test “to the middle” (a crude divide-and-conquer approach to diagnosis E2E problems). Network Monitoring is live (a prototype that will eventually be released) that allows simple analysis of network monitoring data across the backbone. In addition, we’ve made that data available via a web service (conforming to the schemata of the GGF NMWG). Other tools, such as NLANR’s Advisor and the HENP community’s MonALISA tool can now consume that data. 2/25/2019

24 AMI Web Service (W. Matthews) http://abilene. internet2
my %request = ( networkCharacteristic => "path.delay.oneway", subject => { source => { address => { name => "ATLA", }, }, destination => { address => { name => "LOSA", }, }, type => "IPv4", }, startTime => ” T00:00::00-5:00", endTime => ” T23:59:59-5:00", ); my $webService = SOAP::Lite -> uri(' -> -> NetworkMeasurementSet(\%request) -> result; Wrap the request up in an object Send it to the webservice 2/25/2019

25 Test from the Edge to the Middle
Divide and conquer: Partial Path Analysis Install OWAMP and / or BWCTL Obtain a BWCTL key here: Begin testing! 2/25/2019

26 Outreach Establishing Application Communities
HENP, VLBI Establishing Measurement Domains ITECs, CERN, GEANT Collaboration Partners DANTE, ESnet, APAN Standards Integration GGF NMWG, IETF  NLANR / DAST Advisor, MonALISA 2/25/2019

27 Measurement Infrastructure Federation
Why a Federation? Multiple measurement frameworks in existence and under development (piPEs, NLANR Advisor, NLANR AMP, etc.). No static “best practice” measurement framework is likely to emerge, given academics being academics. Future measurement frameworks can build on shoulders of current efforts, not feet. Performance Measurement Architecture Workshop (NSF Grant # ANI ) Matt Zekauskas 2/25/2019

28 Measurement Infrastructure Federation Interfaces
2/25/2019

29 Using piPEs Download, try out BWCTL, OWAMP Test Edge  Middle
Set up Measurement Beacons Identify machines & locations Determine policies and limitations Install BWCTL, OWAMP daemons (Future): Set up measurement domain or application community 2/25/2019

30 Summary We can do partial path analysis, although making sense of the results is still a big issue. We can speak the same measurement language, although it’s still evolving. We are working together in growing numbers, but we need critical mass (become de facto standard). We need to be able to find each other. We need to be able to verify each other’s identity. 2/25/2019

31 2/25/2019

32 Extra Slides 2/25/2019

33 Data Collection / Correlation
Collection Today: Iperf (Throughput) OWAMP (1-Way Latency, Loss) SNMP Data Anonymized Netflow Data Per Sender, Per Receiver, Per Node Pair IPv4 and IPv6 Collection in the Future NTP (Data) Traceroute BGP Data First Mile Analysis Correlation Today: “Worst 10” Throughputs “Worst 10” Latencies Correlation in the Future: 99th Percentile Throughput over Time Throughput/Loss for all E2E paths using a specific link Commonalities among first mile analyzers Sum of Partial Paths vs. Whole Path Throughput is sustained megabits per second across a link. We can do both TCP and UDP versions. (The former is much more useful.) This is a standard measure of TCP performance along a path. Latency is the time it takes a single packet to get from one node to another. Loss is the % of packets that do not make it to their destination (usually 0% in the case of Abilene). SNMP data is gathered on a router and queryable via a protocol (SNMP). Netflow data is passively collected data about traffic through a router. It is anonymized so as to prevent the ability to track a user’s network usage. All this data can be sliced on a per sender, per receiver, or per pair of nodes basis. We are also running IPv4 and IPv6 tests (different protocols for addressing remote computers). NTP data details the accuracy of clocks. (This is very important for evaluating the quality of one-way-latency data, as the clocks need to be synchronized if the data is to mean anything.) Traceroute is a crude form of path detection. BGP data is the right way to determine of routing has changed. First mile analysis is looking for common problems in the first mile (computer to campus edge) where it is believed (but not proved) that most problems occur. Today we can look for the “worst 10” throughputs and latencies. If they are all “good enough” then it’s “not the network”. If there they are not “good enough” then “we have a problem” … crude Network Operations Center (NOC) alarm system. In the future, we expect to be able to look at 99th percentile throughput over time … sliced all the ways mentioned previously … this will give us insights as to the long-term quality of links as traffic generally rises on the network. In the future, we expect to be able to isolate all paths sharing a link in common … see if there’s a network problem common to that link that only manifests over longer paths. (This blurs into the fourth point.) We expect to be able to detect if many users are getting the same first mile problem … perhaps indicative of a pervasive error in configuration within a campus, for example. Some problems only manifest in the “long-haul” … we expect to be able to compare the “long haul” to a succession of “short hauls” and see if a problem manifests. Etc. 2/25/2019

34 Data Analysis Analysis in the Future: Analysis Today: Latency vs. Loss
Throughput over Time Latency over Time Loss over Time Worrisome Tests? (Any bad apples in “Worst Ten”?) “Not the Network” (If “Worst Ten” is good enough) Analysis in the Future: Latency vs. Loss How good is the network? Do common first mile problems exist? Does a link have problems that only manifest in the long-haul? Is the network delivering the performance required by a funded project? First column shows things we can do today in AMD. Second column shows things we could do (i.e. we have the data, but haven’t built the scripts to do the analysis and we’re not sure exactly how to interpret the data yet). 2/25/2019

35 Data Discovery / Interoperability
Discovery in the Future: Where are the measurement nodes corresponding to a specific node? Where are the test results for a specific partial path? Interoperability in the Future: Can I have a test within or to another measurement framework? Can I have a measurement result from within or to another measurement framework? Discovery of measurement nodes and corresponding measurement results is a big, untouched problem. Imagine a “supertraceroute” that could tell you all the nodes (routers) along a path as well as the associated measurement nodes for each router … that’s what we need effectively. Many ways to do it … a need common to lots of groups … no one (as far as we know) yet funded to do such work. 2/25/2019

36 piPEs Deployment APAN 1) Abilene Backbone Deployment (Complete)
Hawaii 2) Hawaii Campus Deployment (Complete) Europe OSU UCSD NC State APAN Abilene Backbone deployment in Green. Campus deployment in Solid Red. In progress roll-out to Campuses and European partners in Dotted Red. “Europe” in this case means a handful of nodes scattered across the GEANT backbone and some nodes at CERN. We are trying to work very closely with DANTE/GEANT and EGEE to ensure interoperability (at a minimum) if not a shared code base across the US and Europe. This is important as many scientists in the US routinely ship data across the Atlantic and thus their E2E concerns are international in nature. 3) Campus, European, APAN Deployment (In Progress) 2/25/2019


Download ppt "Internet2 E2E piPEs Update"

Similar presentations


Ads by Google