Bryan Learn Pittsburgh Supercomputing Center

Slides:



Advertisements
Similar presentations
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
Advertisements

Internet2 and AL2S Eric Boyd Senior Director of Strategic Projects
Title or Title Event/Date Presenter, PresenterTitle, Internet2 Network Virtualization & the Internet2 Innovation Platform To keep our community at the.
Slide title 70 pt CAPITALS Slide subtitle minimum 30 pt Vpn service Ericsson.
XSEDE 13 July 24, Galaxy Team: PSC Team:
Internet2 Network: Convergence of Innovation, SDN, and Cloud Computing Eric Boyd Senior Director of Strategic Projects.
1© Copyright 2015 EMC Corporation. All rights reserved. SDN INTELLIGENT NETWORKING IMPLICATIONS FOR END-TO-END INTERNETWORKING Simone Mangiante Senior.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 8 Introduction to Printers in a Windows Server 2008 Network.
presented by marathonstore.com 2003 Marathon Technology The Western NRG Team Magnum Router: Feature, Function, Benefit Applications Promotions Service.
TeraPaths: A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research Bruce Gibbard & Dantong Yu High-Performance Network Research.
National Science Foundation Arlington, Virginia January 7-8, 2013 Tom Lehman University of Maryland Mid-Atlantic Crossroads.
GN2 Performance Monitoring & Management : AA Needs – Nicolas Simar - 2 nd AA Workshop Nov 2003 Malaga, Spain GN2 Performance Monitoring & Management.
LARK Bringing Distributed High Throughput Computing to the Network Todd Tannenbaum U of Wisconsin-Madison Garhan Attebury
Software-defined Networking Capabilities, Needs in GENI for VMLab ( Prasad Calyam; Sudharsan Rajagopalan;
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
GEC 15 Houston, Texas October 23, 2012 Tom Lehman Xi Yang University of Maryland Mid-Atlantic Crossroads (MAX)
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
TeraPaths TeraPaths: Establishing End-to-End QoS Paths through L2 and L3 WAN Connections Presented by Presented by Dimitrios Katramatos, BNL Dimitrios.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GCRC Meeting 2004 BIRN Coordinating Center Software Development Vicky Rowley.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
EXPOSING OVS STATISTICS FOR Q UANTUM USERS Tomer Shani Advanced Topics in Storage Systems Spring 2013.
1 | © 2015 Infinera Open SDN in Metro P-OTS Networks Sten Nordell CTO Metro Business Group
SOFTWARE DEFINED NETWORKING/OPENFLOW: A PATH TO PROGRAMMABLE NETWORKS April 23, 2012 © Brocade Communications Systems, Inc.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
IPS Infrastructure Technological Overview of Work Done.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
LHCONE NETWORK SERVICES: GETTING SDN TO DEV-OPS IN ATLAS Shawn McKee/Univ. of Michigan LHCONE/LHCOPN Meeting, Taipei, Taiwan March 14th, 2016 March 14,
SDN and Beyond Ghufran Baig Mubashir Adnan Qureshi.
Atrium Router Project Proposal Subhas Mondal, Manoj Nair, Subhash Singh.
UNM SCIENCE DMZ Sean Taylor Senior Network Engineer.
REMOTE MANAGEMENT OF SYSTEM
Md Baitul Al Sadi, Isaac J. Cushman, Lei Chen, Rami J. Haddad
Elastic Cyberinfrastructure for Research Computing
Instructor Materials Chapter 7: Network Evolution
SDN challenges Deployment challenges
Instructor Materials Chapter 1: LAN Design
Business System Development
Multi-layer software defined networking in GÉANT
By: Raza Usmani SaaS, PaaS & TaaS By: Raza Usmani
Software defined networking: Experimental research on QoS
Computing Clusters, Grids and Clouds Globus data service
Grid Optical Burst Switched Networks
University of Maryland College Park
StratusLab Final Periodic Review
StratusLab Final Periodic Review
Bridges and Clouds Sergiu Sanielevici, PSC Director of User Support for Scientific Applications October 12, 2017 © 2017 Pittsburgh Supercomputing Center.
Author: Daniel Guija Alcaraz
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
AP Functional Requirement Initiatives for WLAN Broadcasting
Recap: introduction to e-science
Integration of Network Services Interface version 2 with the JUNOS Space SDK
Enterprise vCPE use case requirement
Introduction to Operating System (OS)
Chapter 18 MobileApp Design
QOS Requirements for Real-Time Services over IP
THE STEPS TO MANAGE THE GRID
SDN Overview for UCAR IT meeting 19-March-2014
GGF15 – Grids and Network Virtualization
XSEDE’s Campus Bridging Project
ONOS Drake Release September 2015.
Managing Clouds with VMM
Software Defined Networking (SDN)
An Introduction to Computer Networking
NSF cloud Chameleon: Phase 2 Networking
ExaO: Software Defined Data Distribution for Exascale Sciences
SISAI STATISTICAL INFORMATION SYSTEMS ARCHITECTURE AND INTEGRATION
Cloud computing mechanisms
Distributed Systems Bina Ramamurthy 4/22/2019 B.Ramamurthy.
Presentation transcript:

CC-NIE DANCES Project: Implementing Multi-Site, SDN-Enabled Applications Bryan Learn <blearn@psc.edu> Pittsburgh Supercomputing Center Internet2 Technology Exchange 26 September 2016

DANCES Project Introduction DANCES is an NSF-funded (#1341005), CC-NIE collaborative “Network Integration and Applied Innovation” project DANCES has integrated SDN/OpenFlow into selected supercomputing cyberinfrastructure applications to support network bandwidth reservation and QoS Motivated by the need to support large bulk file transfer flows and efficiently share bandwidth of end site 10G infrastructure Project was started in Jan2014. The two year project will finish in 2016 under a No Cost Extension Concentrating on CI applications that initiate file transfer. These include: User workflow via scheduling/resource management software (e.g., Torque/Moab, Torque/SLURM) Wide area distributed filesystem (SLASH2) Project was motivated by the need to support large bulk wide area file transfers and the desire to provide predictability and stability throughout the transfer. When the project was proposed, end site components typically had 10Gb/s interfaces. Internet2 had recently deployed 100Gb/s uncongested SDN-capable infrastructure in the wide area. Multiple 10G attached servers at end sites could cause local congestion while the WAN path remained uncongested. Large file transfers can take days to complete. Depending on how smaller competing file transfers are initiated, they may either disrupt the larger flow, or the larger flow may block or seriously impede performance of smaller flows. Through the DANCES project we are trying to bring more predictability to file transfer performance by providing dedicated and protected bandwidth for priority bulk data flows.

DANCES Development Team Pittsburgh Supercomputing Center (PSC) National Institute for Computational Sciences (NICS) Pennsylvania State University (Penn State) eXtreme Science and Engineering Discovery Environment (XSEDE) Internet2 Corsa Technologies, Inc. Our core team

DANCES-enhanced Applications Implementation uses SDN/OpenFlow 1.3 metering to implement QoS The enhanced cyberinfrastructure applications are: Supercomputing resource management and scheduling. File transfer via GridFTP scp can then be scheduled and included as part of a workflow Integrated into a distributed wide area file system SLASH2 The metering capability and marking of packets, along with port queueing and weighting functionality, implements QoS SLASH2 was developed by PSC and is an open source wide area network (WAN)-friendly distributed file system offering advanced features such as data multi-residency, system-managed data transfer, and inline checksum verification. We are working to incorporate the network bandwidth scheduling capability with SLASH2 file replication.

SDN/OF Infrastructure Components and Interfaces CONGA bandwidth manager software Receives bandwidth request from application Verifies user authorization with user database Tracks available/allocated bandwidth between sites Initiates OpenFlow path set up via the RYU OpenFlow controller RYU OpenFlow controller software Writes flowmods (routing rules) to switches Reads port statistics to verify operation OpenFlow 1.3-compatible switch hardware Expand as you see fit

The next few slides step through the control and data flow of the system. User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

Workflow Example User submits job (e.g., via qsub) that requires file transfer with reserved bandwidth QoS. Bandwidth request is initiated in Torque Prologue CONGA checks user authorization in XSEDE User Database If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth This slide and the next are text for the previous diagrams so may or may not be useful to include. User submits job (via qsub, for example) that requires file transfer with reserved bandwidth QoS (can also include compute and storage requests) to the Resource Manager / Scheduler. Torque Prologue script initiates RM/S check with CONGA for user auth and available bandwidth between src and dst. CONGA checks user authorization listed in XSEDE User Database for bandwidth scheduling (We worked with the XSEDE allocations/database team and they created a small database with entries for the DANCES project team members. The test db is accessible via REST API) and determines [(based on tracking state and/or polling?)] if bandwidth is available. If not, [best effort and inform user?] If bandwidth is available, CONGA will initiate path and bandwidth provisioning with the RYU OF controller. RYU OF controller provisions local and remote flow and bandwidth

Workflow Example Torque/<scheduler> schedules job when resources are available. System has been tested with Moab local scheduling at NICS and PBS at PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished Torque/<scheduler> schedules job when resources are available. System has been used with Moab@NICS and PBS@PSC; PSC in production will likely use SLURM Execute file transfer Torque Epilogue script or timeout will initiate tear down of provisioned path when transfer is finished

Implementation CONGA RYU OpenFlow controller Custom software created as the “Northbound API” for DANCES project Uses REST API RYU OpenFlow controller Production deployment will require extension to XSEDE allocation system to accommodate a dedicated network bandwidth resource to be managed along with traditional compute and storage resources Accounting by duration of dedicated bandwidth request The implementation of DANCES required significant software development and integration effort. CONGA: Created a “northbound API” for CONGA to enable the applications (Torque and SLASH2) to communicate with the RYU OpenFlow controller framework. With the concept of QoS request scheduling and cross domain operation, CONGA also needed to track bandwidth allocation. REST API chosen to ease extensibility RYU Provided as an OpenFlow controller “framework” with example modules Customize for specific applications Integration with existing XSEDE operational model Using a test database of DANCES development team members in XSEDE user database format

Implementation and Deployment Original plan was full interoperation with Internet2’s AL2S for end-to-end SDN Internet2’s upcoming transition to MPLS removes SDN from WAN production infrastructure to a 10Gb research overlay The DANCES components are still applicable to campuses with congested 10Gb connectivity and DTNs close to the edge Each site would run an instance of CONGA or some other type of resource manager/scheduler and RYU to locally manage devices and congestion Created software module to interface with OESS to provision VLANs on demand. B/W provisioning was not supported by OF1.0 on AL2S, but we tested QoS enforced at end point switches. Will be working with I2 to define interoperation as I2 transitions to MPLS (Someone from I2 may add comments) With the campus to XSEDE SP concept, resource contention is more likely at the campus, not within the WAN path or SP so DANCES functionality would improve throughput without requiring support at both ends or completely end-to-end

Observations Cross-domain control of resources is challenging SDN/OF has been slower to gain WAN deployment support than originally expected (or has gone the opposite direction) Control of multi-vendor SDN environments requires additional custom coding By Bullet2 I’m referring to I2 stepping away from SDN/OF to MPLS

Future Work Additional testing of OpenFlow 1.3 flow metering with various queue sizes and weights Measure network bandwidth utilization achieved by using bandwidth scheduling for large flows along with “best effort” traffic Expand test deployment Prepare and package for production deployment Project is drawing to a close, but in the time remaining, we may...

Questions? https://www.dances-sdn.org dances-dev@dances-sdn.org