DYNES ( DYnamic NEtwork System) & LHCONE ( LHC Open Network Env.) Shawn McKee University of Michigan Jason Zurawski Internet2 USATLAS Facilities Meeting.

Slides:



Advertisements
Similar presentations
Circuit Monitoring July 16 th 2011, OGF 32: NMC-WG Jason Zurawski, Internet2 Research Liaison.
Advertisements

High Performance Computing Course Notes Grid Computing.
Internet2 and AL2S Eric Boyd Senior Director of Strategic Projects
Title or Title Event/Date Presenter, PresenterTitle, Internet2 Network Virtualization & the Internet2 Innovation Platform To keep our community at the.
America Connects to Europe (ACE) and TransPAC3 (TP3) Cooperative Partnerships to facilitate Global R/E Collaboration John Hicks – ACE and TP3 Engineer.
Kansei Connectivity Requirements: Campus Deployment Case Study Anish Arora/Wenjie Zeng, GENI Kansei Project Prasad Calyam, Ohio Supercomputer Center/OARnet.
GNEW 2004 CERN, Geneva, Switzerland March 16th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview for GNEW2004 Shawn McKee University of Michigan.
Integrating Network and Transfer Metrics to Optimize Transfer Efficiency and Experiment Workflows Shawn McKee, Marian Babik for the WLCG Network and Transfer.
Trial of the Infinera PXM Guy Roberts, Mian Usman.
Institute of Technology, Sligo Dept of Computing Semester 3, version Semester 3 Chapter 3 VLANs.
Transatlantic Connectivity in LHCONE Artur Barczyk California Institute of Technology LHCONE Meeting Washington DC, June 13 th,
ESnet On-demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok Network Engineering Group Thomas Ndousse Visit February Energy.
TeraPaths : A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research USATLAS Tier 1 & Tier 2 Network Planning Meeting December.
IRNC Special Projects: IRIS and DyGIR Eric Boyd, Internet2 October 5, 2011.
1 ESnet Network Measurements ESCC Feb Joe Metzger
Circuit Services - IPTV Christian Todorov Internet2 Fall Member Meeting October 9, 2007.
Virtual LAN Design Switches also have enabled the creation of Virtual LANs (VLANs). VLANs provide greater opportunities to manage the flow of traffic on.
TeraPaths: A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research Bruce Gibbard & Dantong Yu High-Performance Network Research.
National Science Foundation Arlington, Virginia January 7-8, 2013 Tom Lehman University of Maryland Mid-Atlantic Crossroads.
Fermi National Accelerator Laboratory, U.S.A. Brookhaven National Laboratory, U.S.A, Karlsruhe Institute of Technology, Germany CHEP 2012, New York, U.S.A.
The DYNES Instrument: A Description and Overview May 24 th 2012 – CHEP 2012 Jason Zurawski, Senior Network Engineer - Internet2.
TeraPaths TeraPaths: establishing end-to-end QoS paths - the user perspective Presented by Presented by Dimitrios Katramatos, BNL Dimitrios Katramatos,
GEC 15 Houston, Texas October 23, 2012 Tom Lehman Xi Yang University of Maryland Mid-Atlantic Crossroads (MAX)
Internet2 Performance Update Jeff W. Boote Senior Network Software Engineer Internet2.
InterDomain Dynamic Circuit Network Demo Joint Techs - Hawaii Jan 2008 John Vollbrecht, Internet2
LHC Open Network Environment LHCONE Artur Barczyk California Institute of Technology LISHEP Workshop on the LHC Rio de Janeiro, July 9 th,
1 Measuring Circuit Based Networks Joint Techs Feb Joe Metzger
Thoughts on Future LHCOPN Some ideas Artur Barczyk, Vancouver, 31/08/09.
A Framework for Internetworking Heterogeneous High-Performance Networks via GMPLS and Web Services Xi Yang, Tom Lehman Information Sciences Institute (ISI)
Delivering Circuit Services to Researchers: The HOPI Testbed Rick Summerhill Director, Network Research, Architecture, and Technologies, Internet2 Joint.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
High Energy & Nuclear Physics (HENP) SIG October 4 th 2011 – Fall Member Meeting Jason Zurawski, Internet2 Research Liaison.
LHC Open Network Environment LHCONE David Foster CERN IT LCG OB 30th September
Scientific Networking: The Cause of and Solution to All Problems April 14 th Workshop on High Performance Applications of Cloud and Grid Tools Jason.
Connect communicate collaborate GÉANT3 Services Connectivity and Monitoring Services by and for NRENs Ann Harding, SWITCH TNC 2010.
ASCR/ESnet Network Requirements an Internet2 Perspective 2009 ASCR/ESnet Network Requirements Workshop April 15/16, 2009 Richard Carlson -- Internet2.
TeraPaths TeraPaths: Establishing End-to-End QoS Paths through L2 and L3 WAN Connections Presented by Presented by Dimitrios Katramatos, BNL Dimitrios.
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
Connect. Communicate. Collaborate perfSONAR MDM Service for LHC OPN Loukik Kudarimoti DANTE.
DYnamic NEtwork System (DYNES) NSF # October 3 rd 2011 – Fall Member Meeting Eric Boyd, Internet2 Jason Zurawski, Internet2.
Internet2 Update October 7 th 2010, LHCOPN Jason Zurawski, Internet2.
Cisco S3C3 Virtual LANS. Why VLANs? You can define groupings of workstations even if separated by switches and on different LAN segments –They are one.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
NORDUnet Nordic Infrastructure for Research & Education Workshop Introduction - Finding the Match Lars Fischer LHCONE Workshop CERN, December 2012.
LHC Open Network Environment Architecture Overview and Status Artur Barczyk/Caltech LHCONE meeting Amsterdam, September 26 th,
Connect communicate collaborate LHCONE moving forward Roberto Sabatino, Mian Usman DANTE LHCONE technical workshop SARA, 1-2 December 2011.
DYNES Project Updates October 11 th 2011 – USATLAS Facilities Meeting Shawn McKee, University of Michigan Jason Zurawski, Internet2.
Dynamic Circuit Network An Introduction John Vollbrecht, Internet2 May 26, 2008.
SDN and OSCARS how-to Evangelos Chaniotakis Network Engineering Group ESCC Indianapoilis, July 2009 Energy Sciences Network Lawrence Berkeley National.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
Internet2 Dynamic Circuit Services and Tools Andrew Lake, Internet2 July 15, 2007 JointTechs, Batavia, IL.
ANSE: Advanced Network Services for Experiments Institutes: –Caltech (PI: H. Newman, Co-PI: A. Barczyk) –University of Michigan (Co-PI: S. McKee) –Vanderbilt.
DICE: Authorizing Dynamic Networks for VOs Jeff W. Boote Senior Network Software Engineer, Internet2 Cándido Rodríguez Montes RedIRIS TNC2009 Malaga, Spain.
LHCONE Monitoring Thoughts June 14 th, LHCOPN/LHCONE Meeting Jason Zurawski – Research Liaison.
Connect communicate collaborate LHCONE European design & implementation Roberto Sabatino, DANTE LHCONE Meeting, Washington, June
Strawman LHCONE Point to Point Experiment Plan LHCONE meeting Paris, June 17-18, 2013.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
LHCONE NETWORK SERVICES: GETTING SDN TO DEV-OPS IN ATLAS Shawn McKee/Univ. of Michigan LHCONE/LHCOPN Meeting, Taipei, Taiwan March 14th, 2016 March 14,
The DYNES Architecture & LHC Data Movement Shawn McKee/University of Michigan For the DYNES collaboration Contributions from Artur Barczyk, Eric Boyd,
LHCONE NETWORK SERVICES INTERFACE (NSI) POINT-TO-POINT TESTBED WITH ATLAS SITES Shawn McKee/Univ. of Michigan Kaushik De/Univ. of Texas Arlington (Thanks.
1 Network Measurement Challenges LHC E2E Network Research Meeting October 25 th 2006 Joe Metzger Version 1.1.
LHCOPN / LHCONE Status Update John Shade /CERN IT-CS Summary of the LHCOPN/LHCONE meeting in Amsterdam Grid Deployment Board, October 2011.
The Internet2 Network and LHC Rick Summerhill Director Network Research, Architecture, and Technologies Internet2 LHC Meeting 23 October 2006 FERMI Lab,
InterDomain Dynamic Circuit Network Demo
Robert Szuman – Poznań Supercomputing and Networking Center, Poland
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
Tailor slide to customer industry/pain points
Big-Data around the world
Chapter 3 VLANs Chaffee County Academy
The UltraLight Program
Presentation transcript:

DYNES ( DYnamic NEtwork System) & LHCONE ( LHC Open Network Env.) Shawn McKee University of Michigan Jason Zurawski Internet2 USATLAS Facilities Meeting October 11, 2011 Southern Methodist University Dallas, Texas

Overview/Outline I want to quickly review DYNES and LHCONE so everyone understands what they are, how they are related and be able to consider how we (USATLAS) might take advantage of them. Then I want to raise some discussion points ◦ DYNES integration within the facility…how? ◦ LHCONE participation and planning…when? ◦ DYNES-LHCONE interactions…why? 2

DYNES Summary NSF MRI-R2: DYnamic NEtwork System (DYNES, NSF # ) NSF MRI-R2: DYnamic NEtwork System (DYNES, NSF # ) What is it?: ◦ A nationwide cyber-instrument spanning up to ~40 US universities and ~14 Internet2 connectors  Extends Internet2s ION service into regional networks and campuses, based on OSCARS implementation of IDC protocol (developed in partnership with ESnet) Who is it? ◦ A collaborative team including Internet2, Caltech, University of Michigan, and Vanderbilt University ◦ Community of regional networks and campuses ◦ LHC, astrophysics community, OSG, WLCG, other virtual organizations What are the goals? ◦ Support large, long-distance scientific data flows in the LHC, other leading programs in data intensive science (such as LIGO, Virtual Observatory, and other large scale sky surveys), and the broader scientific community ◦ Build a distributed virtual instrument at sites of interest to the LHC but available to R&E community generally 3

LHCONE Summary 4 LHCONE LHCONE - LHC Open Network Environment Results of LHC Tier-2 network working group convened summer A merger of 4 “whitepapers” from the CERN LHCT2 meeting in January 2011 (See ) LHCONE builds on the hybrid network infrastructures and open exchange points provided today by the major R&E networks on all continents Goal: To build a global unified service platform for the LHC community By design, LHCONE makes best use of the technologies and best current practices and facilities provided today in national, regional and international R&E network

Solutions: ◦ Dedicated bandwidth (over the entire end-to-end path) to move scientific data ◦ Invoke this “on demand” instead of relying on permanent capacity (cost, complexity) ◦ (Co-)Exists in harmony with traditional IP networking ◦ Connect between facilities that scientists need to access ◦ Integration with data movement applications  Invoke the connectivity when the need it, based on network conditions Prior Work ◦ “Dynamic Circuit” Networking – creation of Layer 2 point to point VLANs ◦ Transit the Campus, Regional, and Backbone R&E networks ◦ Software to manage the scheduling and negotiation of resources DYNES Concepts

DYNES: Why Not Static Circuits or Traditional, General Purpose Networks ? Separation (physical or logical) of the dynamic circuit-oriented network from the IP backbone is driven by the need to meet different functional, security, and architectural needs: ◦ Static “nailed-up” circuits will not scale. ◦ GP network firewalls incompatible with enabling large-scale science network dataflows ◦ Implementing many high capacity ports on traditional routers would be very expensive  Price balance: Worse in the next generation: 40G and 100G general purpose router ports are several hundred k$ each.

Initial Deployment Locations: ◦ 30 End Sites ◦ 8 Regional Networks ◦ Collaboration with like minded efforts (DoE ESCPS and StorNet) Accepting additional applications ◦ Supporting all data-intensive, distributed science - e arly focus on Physics (LHC) sites DYNES Scope

DYNES Infrastructure Overview DYNES Topology ◦ Based on Applications received ◦ Plus existing peering wide area Dynamic Circuit Connections (DCN ) 8

Inter-domain Controller (IDC) Server and Software ◦ IDC creates virtual LANs (VLANs) dynamically between the FDT server, local campus, and wide area network ◦ Dell R410 (1U) Server ◦ OSCARSv0.6 and DRAGON Software Fast Data Transfer (FDT) server ◦ Fast Data Transfer (FDT) server connects to the disk array via the SAS controller and runs the FDT software ◦ Dell R510 (2U) Server DYNES Ethernet switch options: ◦ Dell PC6248 (48 1GE ports, 4 10GE capable ports (SFP+, CX4 or optical) ◦ Dell PC8024F (24 10GE SFP+ ports, 4 “combo” ports supporting RJ45; SFP+ supporting optical) perfSONAR Monitoring DYNES Standard Equipment and Software Many thanks to Dell for supporting DYNES with LHC pricing!

DYNES Data Flow Overview USATLAS Site Z Resources USATLAS Site A Resources ???

4 Project Phases ◦ Phase 1: Site Selection and Planning ( Completed in Feb 2011 ) ◦ Phase 2: Initial Devel./Deployment ( Feb 2011 through July 2011 ) ◦ Phase 3: Scale up to Full Deployment ( July 2011 through Aug 2012 )  DYNES Participant Deployment (July 2011-November 2011)  Full-scale System Development, Testing & Evaluation (November 2011 – August 2012) ◦ Phase 4: Integration at Scale; Transition to Routine O&M ( Aug 2012 through August 2013 ) ◦ Details in supplemental slides at end of this talk A DYNES Program Plan document along with many other documents are available at: ◦ Questions can be sent to the mailing list: ◦ DYNES Current Status

The LHCONE effort has been moving ahead via meetings during 2011 GOALS: Identify, organize and manage LHC related flows from the Tier-2 and Tier-3 sites Having a way to identify LHC related flows helps by: ◦ Allowing this traffic to be “engineered” by whatever means exist within the infrastructure ◦ Makes monitoring much more straightforward ◦ Enables quicker problem isolation and resolution ◦ Motivates additional resources for LHC needs How to get LHCONE moving? LHCONE Goals

“Joe’s Solution” – Result of June 2011 Meeting Two “issues” identified at the DC meeting as needing particular attention: Multiple paths across Atlantic Resiliency Agreed to have the architecture group work out a solution Layer 2 ‘islands’ joined by Layer 3 connections

Multipoint: ◦ Domains interconnected through Layer 2 switches ◦ Two VLANs (nominal IDs: 3000, 2000)  VLAN 2000 configured on GEANT/ACE transatlantic segment  VLAN 3000 configured on US LHCNet transatlantic segment ◦ Allows to use both TA segments, provides TA resiliency ◦ 2 route servers per VLAN  Each connecting site peers will all 4 route servers ◦ Enables up to 25G on the Trans-Atlantic routes for LHC traffic. Point to Point: ◦ Suggestion: Build on efforts of DYNES and DICE-Dynamic service ◦ DICE-Dynamic service being rolled out by ESnet, GÉANT, Internet2, and USLHCnet  Remaining issues being worked out  Planned commencement of service: October, 2011  Built on OSCARS (ESnet, Internet2, USLHCnet, RNP) and AutoBAHN (GÉANT), using IDC protocol LHCONE Pilot

LHCONE Interim Solution ATLANTIC OCEAN Slide A. Barczyk

LHCONE in Starlight (Interim) 16 Slide A. Barczyk

Continents Connected? ◦ Access Switches in the US and Europe, work to get Asia connected ◦ VLANs stretched about as far as they can stretch ◦ Capacity identified (e.g. NA, EU and TA) – but note that in many cases this is not as much as would be traditionally available (e.g. it’s a pilot built from ‘spare’ parts) ◦ Basic functionality in place, end sites are joining – can success be measured? Technology “Sound”? ◦ Lots of questions about this, Layer2 Islands w/ Layer3 connections is hard to manage (and debug) ◦ Already routing loop situations (e.g. the ‘2 VLAN Solution’ is not really a sound solution) ◦ Serious doubts about what happens as more TA links get added – more sites in the US/EU come online with multiple access locations (e.g. the situatuon in the Chicago region is hard to manage…) LHCONE Pilot … But Where Are We Really?

Customers Happy? ◦ A very relevant and timely question asked by Michael at the SARA Meeting ◦ Oversight is lacking ◦ Not a lot of information being shared with the ‘stakeholders’. How to fix? Architectural Changes? ◦ Some proposals to scrap the whole thing and start over ◦ Some questions about if the scalability issues we see today will cloud the overall concept It is critical that we stay engaged in LHCONE while realizing the current limitations implied by building something from donated/reused components and effort. LHCONE Pilot … But Where Are We Really?

LHCONE MWT2_UC L2 Example 19 Slide A. Nickolich B. O’Keefe

LHCONE MWT2_UC L3 Example 20 Slide A. Nickolich B. O’Keefe

LHCONE Planned “Features” LHCONE should be a well defined architecture once it is put into production It has always included the following concepts: ◦ Multipoint VLAN (extent is under debate) ◦ Point-to-Point connections ◦ A routed IP component (extent is under debate) ◦ Traffic-engineering (L1, L2 and L3 options) 21

LHCONE Planning for USATLAS For USATLAS, we need to determine how best to participate within the LHCONE effort. Practically, this means how and when our sites might enable LHCONE. We need to have well-defined metrics in place which we gather before and after adding a site to LHCONE We can discuss this in a bit… 22

DYNES Integration in USATLAS One goal of DYNES is getting the needed network capabilities pushed out to user-sites (last-mile). This helps everyone to be able to utilize the new DCN capabilities. Tier-3 Tier-2sTier-3 A natural primary use of DYNES for USATLAS is to allow Tier-3 sites to prioritize data flows as needed, primarily from Tier-2s or other Tier-3s. ◦ We need to start testing how we can integrate DYNEs capabilities semi-transparently for USATLAS users. (First steps…) ◦ Goal is to improve the end-user experience in getting larger amounts of data in a timely way compared to the current situation. ◦ Eventually should be completely transparent and integrated with our tools such that DYNES sites automatically utilize DYNES.

DYNES and LHCONE: DYNES as an “On-ramp” to LHCONE DYNES Participants DYNES Participants can dynamically connect to Exchange Points via ION Service DYNES Participants can dynamically connect to Exchange Points via ION Service DYNES operation is currently based on and "end- to-end" model, so will need to make some adjustment with respect to how to extend services/connections thru Exchange Point to service endpoints DYNES operation is currently based on and "end- to-end" model, so will need to make some adjustment with respect to how to extend services/connections thru Exchange Point to service endpoints ◦ Dynamic Circuits thru and beyond the exchange point? ◦ Hybrid dynamic circuit and IP routed segment model?

Other Projects Leveraging Dynamic Circuits at the HEP sites: StorNet, ESCPS StorNet – BNL, LBNL, UMICH ◦ Integrated Dynamic Storage and Network Resource Provisioning and Management for Automated Data Transfers ESCPS – FNAL, BNL, Delaware ◦ End Site Control Plane System 25 Building on previous developments in and experience from the TeraPaths and LambdaStation projects StorNet: Integration of TeraPaths and BeStMan

Integration into ATLAS/CMS software stack ◦ Started a thread with PhEDEx developers, Shawn/Michael/Jason had a minor thread about DQ2 ◦ Should the ‘intelligent’ data movers know there is a network choice? Should the low level tool only be aware? Enabling at more T3s ◦ More than in our current set. We will most likely have more funding available in the end – candidates? Reaching Europe, Asia, South America (Australia? Africa?) ◦ Can plug into RNP/GEANT pretty easily. ◦ APAN has OSCARS working to a certain extent in Asia ◦ IRNC (NSF) international links have to support OSCARS, this gets the tech into South America/India/Australia/Asia/Europe, etc. ◦ What is most important, and should be pursued? LHCONE ◦ Lots more questions in this area than answers…  One positive thing is that DYNES will enable dynamic access into VLANs … but will LHCONE still be pushing the concepts discussed this year? ◦ Is all the work beneficial to experiments? Open Questions

DYNES/LHCONE Discussion(1) How to start the DYNES integration into USATLAS? ◦ Requires a deployed “instrument” (Feb 2012?) ◦ Need to think about ATLAS DDM, Xrootd Federation, standard tools and how best to connect with DYNES…can influence/modify DYNES “design” ◦ Familiarize users with DYNES first, then begin “integration”? Or start out now with a subset of the facility? 27

DYNES/LHCONE Discussion(2) How best to test and influence LHCONE? ◦ Current solution is temporary and perhaps not “production” level. ◦ Testing for Tier-2’s will be interesting to verify:  Performance doesn’t decrease  Problems (esp. trans-oceans transfers) are less visible and/or easier to debug/fix ◦ Suggest we define a specific timeline to have AGLT2 and/or MWT2 test LHCONE as is.  Benchmarks for current setup?  Comparison after transition to LHCONE participation 28

Summary and Other Discussion LHCONE and DYNES are getting close to being useable for USATLAS. There are other resource projects that may prove useful once we have circuits. We need to be planning for how to utilize these capabilities for the facility ◦ Our input can help ensure they will meet our needs Other issues we should cover? Questions? 29

DYNES/LHCONE References DYNES ◦ LHCONE ◦ OSCARS ◦ DRAGON ◦ DCN Software Suite (DCNSS) ◦ FDT ◦

ADDITIONAL SLIDES DYNES and LHCONE supplemental information 31

DYNES Demo Before Jason’s talk let’s see a quick example of how DYNES can work… 32

Inter-domain Controller (IDC) Server and Software – IDC creates virtual LANs (VLANs) dynamically between the FDT server, local campus, and wide area network – IDC software is based on the OSCARS and DRAGON software which is packaged together as the DCN Software Suite (DCNSS) – DCNSS version correlates to stable tested versions of OSCARS. The current version of DCNSS is v – Initial DYNES deployments will include both DCNSSv0.6 and DCNSSv0.5.4 virtual machines Currently XEN based Looking into KVM for future releases DYNES IDC

The DYNES Agent (DA) will provide the functionality to request the circuit instantiation, initiate and manage the data transfer, and terminate the dynamically provisioned resources. Specifically the DA will do the following: ◦ Accept user request in the form of a DYNES Transfer URLs indicating the data location and ID ◦ Locates the remote side DYNES EndPoint Name embedded in the Transfer URL ◦ Submits a dynamic circuit request to its home InterDomain Controller (IDC) utilizing its local DYNES EndPoint Name as source and DYNES EndPoint Name from Transfer URL as the destination ◦ Wait for confirmation that dynamic circuit has been established ◦ Starts and manages Data Transfer using the appropriate DYNES Project IP addresses ◦ Initiate release of dynamic circuit upon completion DYNES FDT