SENSE: SDN for End-to-end Networked Science at the Exascale

Slides:



Advertisements
Similar presentations
LambdaStation Phil DeMar Don Petravick NeSC Oct. 7, 2004.
Advertisements

Application-Based Network Operations (ABNO) IETF 88 – SDN RG
Lawrence Berkeley National LaboratoryU.S. Department of Energy | Office of Science Network Service Interface (NSI) Inder Monga Co-chair, Network Services.
© 2006 Open Grid Forum Network Service Interface in a Nut Shell GEC 19, Atlanta, GA Presenter: Chin Guok (ESnet) Contributors: Tomohiro Kudoh (AIST), John.
DRAGON Dynamic Resource Allocation via GMPLS Optical Networks Tom Lehman University of Southern California Information Sciences Institute (USC/ISI) National.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
4.1.5 System Management Background What is in System Management Resource control and scheduling Booting, reconfiguration, defining limits for resource.
Grant agreement n° SDN architectures for orchestration of mobile cloud services with converged control of wireless access and optical transport network.
SmartER Semantic Cloud Sevices Karuna P Joshi University of Maryland, Baltimore County Advisors: Dr. Tim Finin, Dr. Yelena Yesha.
CLAG 2004 – April/041 A Workflow-based Architecture for e- Learning in the Grid Luiz A. Pereira, Fábio A. Porto, Bruno Schulze, Rubens N. Melo
Demonstrating IT Relevance to Business Aligning IT and Business Goals with On Demand Automation Solutions Robert LeBlanc General Manager Tivoli Software.
Extreme Networks Confidential and Proprietary. © 2010 Extreme Networks Inc. All rights reserved.
©2013 Avaya Inc. All rights reservedFebruary 26-28, 2013 | Orlando, FL.
Use Case for Distributed Data Center in SUPA
National Science Foundation Arlington, Virginia January 7-8, 2013 Tom Lehman University of Maryland Mid-Atlantic Crossroads.
GN2 Performance Monitoring & Management : AA Needs – Nicolas Simar - 2 nd AA Workshop Nov 2003 Malaga, Spain GN2 Performance Monitoring & Management.
DRAGON Dynamic Resource Allocation via GMPLS Optical Networks API Overview Jaroslav Flidr, Peter O’Neil, Chris Tracy Mid-Atlantic Crossroads.
GEC 15 Houston, Texas October 23, 2012 Tom Lehman Xi Yang University of Maryland Mid-Atlantic Crossroads (MAX)
© 2012 xtUML.org Bill Chown – Mentor Graphics Model Driven Engineering.
Vic Liu Liang Xia Zu Qiang Speaker: Vic Liu China Mobile Network as a Service Architecture draft-liu-nvo3-naas-arch-01.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
SDN AND OPENFLOW SPECIFICATION SPEAKER: HSUAN-LING WENG DATE: 2014/11/18.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
LHC OPEN NETWORK ENVIRONMENT STATUS UPDATE Artur Barczyk/Caltech Tokyo, May 2013 May 14, 2013
SDN Management Layer DESIGN REQUIREMENTS AND FUTURE DIRECTION NO OF SLIDES : 26 1.
GRID ANATOMY Advanced Computing Concepts – Dr. Emmanuel Pilli.
D. Stiliadis F. Balus W. Henderickx N. Bitar M. Pisica Software Driven Networks: Use Cases and Framework draft-stiliadis-sdnp-framework-use-cases-01.txt.
Cyberinfrastructure Overview of Demos Townsville, AU 28 – 31 March 2006 CREON/GLEON.
Supporting Advanced Scientific Computing Research Basic Energy Sciences Biological and Environmental Research Fusion Energy Sciences High Energy Physics.
ESnet’s Use of OpenFlow To Facilitate Science Data Mobility Chin Guok Inder Monga, and Eric Pouyoul OGF 36 OpenFlow Workshop Chicago, Il Oct 8, 2012.
Presenter: Prof. Dimitris Mourtzis Advanced Manufacturing: Industry 4.0 and Smart Systems.
DATA ACCESS and DATA MANAGEMENT CHALLENGES in CMS.
Issues in Cloud Computing. Agenda Issues in Inter-cloud, environments  QoS, Monitoirng Load balancing  Dynamic configuration  Resource optimization.
Software Defined Networking BY RAVI NAMBOORI. Overview  Origins of SDN.  What is SDN ?  Original Definition of SDN.  What = Why We need SDN ?  Conclusion.
GENI Enabled Software Defined Exchange (SDX) and ScienceDMZ (SD-SDMZ)
OpenDaylight Based Machine Learning for Networks
Defining ONAP APIs With BSS/OSS
Service Assurance in the Age of Virtualization
Lifecycle Service Orchestration (LSO) Models in context
Organizations Are Embracing New Opportunities
Multi-layer software defined networking in GÉANT
Bryan Learn Pittsburgh Supercomputing Center
Use Case for Distributed Data Center in SUPA
Grid Optical Burst Switched Networks
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Heitor Moraes, Marcos Vieira, Italo Cunha, Dorgival Guedes
Effective Network Orchestration Starts by Automating Provisioning Downloadable Figures Simon Richard.
MEF Modeling Activities
GÉANT Multi-Domain Bandwidth-on-Demand Service
DOE Facilities - Drivers for Science: Experimental and Simulation Data
Introduction to Data Management in EGI
Author: Daniel Guija Alcaraz
Joint Genome Institute
Grid Network Services: Lessons from SC04 draft-ggf-bas-sc04demo-0.doc
Integration of Network Services Interface version 2 with the JUNOS Space SDK
Regional Software Defined Science DMZ (SD-SDMZ)
Chapter 18 MobileApp Design
Street Cleanliness Assessment System for Smart City using Mobile and Cloud Bharat Bhushan, Kavin Pradeep Sriram Kumar, Mithra Desinguraj, Sonal Gupta Project.
Policy based Cloud Services on a VCL platform
Borderless Networks Services Supported
Continuous Automated Chatbot Testing
PRPv1 Discussion topics
ExaO: Software Defined Data Distribution for Exascale Sciences
Core Platform The base of EmpFinesse™ Suite.
See your OpenStack Network Like Never Before
OSCARS Roadmap Chin Guok
Best practises and experiences in user support
Utilizing the Network Edge
Transport network requirements for TAPI considering NBI
Presentation transcript:

SENSE: SDN for End-to-end Networked Science at the Exascale The International Conference for High Performance Computing, Networking, Storage and Analysis November 12-17, 2017 Denver, Colorado SENSE Team Sponsor Advanced Scientific Computing Research (ASCR)

SENSE Team Harvey Newman Justas Balcas Maria Spiropulu Tom Lehman ESnet Inder Monga Chin Guok John MacAuley Fermilab Phil Demar ANL Linda Winkler NERSC Damian Hazen Caltech Harvey Newman Justas Balcas Maria Spiropulu University of Maryland/Mid-Atlantic Crossroads Tom Lehman Xi Yang

What Research Problem(s) are We Solving? Distributed scientific workflows need end-to-end automation so the focus can be on science, and not infrastructure Manual provisioning and infrastructure debugging takes time No service consistency across domains No service visibility or automated troubleshooting across domains Lack of realtime information from domains impedes development of intelligent services

What Research Problem(s) are We Solving? Science application workflows need to drive network service provisioning Network programming APIs usually not intuitive and require detailed network knowledge, some not pre-known Detailed network information needed, usually not easily available Efficient use of resources Multi-domain orchestration requires service visibility and troubleshooting Data APIs across domains for applications, users, network administrators Performance, service statistics, topology, capability, etc. Exchange of ‘scoped’ and authorized information

Vision and Objectives A new paradigm for Application to Network Interactions Intent Based – Abstract requests and questions in the context of the application objectives. Interactive – what is possible? what is recommended? let’s negotiate. Real-time – resource availability, provisioning options, service status, troubleshooting. End-to-End – multi-domain networks, end sites, and the network stack inside the end systems. Full Service Lifecycle Interactions – continuous conversation between application and network for the service duration.

Application Workflow Agent SENSE Architecture Application Workflow Agent Model Driven SDN Control with Orchestration Intent Based APIs with Resource Discovery, Negotiation, Service Lifecycle Monitoring/Troubleshooting SENSE End-to-End Model Datafication of cyberinfrastructure to enable intelligent services SENSE Orchestrator Realtime system based on Resource Manager developed infrastructure and service models RM DTN [200G] RM DTN [40G] RM DTN [40G] RM DTN [50G] UMD RM RM RM ESnet MAX DTN [200G] RM SENSE Resource Manager

SENSE Orchestrator API Application Workflow Agent Resource Discovery Service Service Discovery End-Point Listing Connectivity Service Point-to-Point Multi-Point Resource Computation Service Monitoring and Troubleshooting Service Types of Interactions What is possible? What is recommended? Requests with negotiation Service status and troubleshooting SENSE Orchestrator RM DTN [200G] RM DTN [40G] RM RM DTN [40G] DTN [50G] UMD RM RM RM DTN [200G] ESnet MAX

SENSE Resource Manager API Allow the machines to automate, iterate, react, and adjust to find solutions and not bring the humans in until absolutely necessary Model Based Control and Orchestration SENSE Orchestrator Control Learn system states Read/Sync data Network Markup Language/RDF based Model Schema Change system states Write / Sync data Multi-Resource Modeling Feedback / Awareness Automatic Operation Infrastructure RM DTN [200G] RM RM DTN [40G] RM DTN [40G] DTN [50G] UMD RM RM RM MAX DTN [200G] ESnet SENSE Resource Manager

Application to SENSE Interactions Please provide a listing of all available provisioning Endpoints Application Workflow Agent Request 20 Gbps P2P service between Caltech and Fermilab. If 20 Gbps not available 10Gbps is ok. What is the maximum bandwidth available for a P2P service between Caltech and Fermilab? Please provide a listing of all available provisioning Endpoints Service is not working, please check status Endpoint Listing What is the maximum bandwidth available for a P2P service between Caltech and Fermilab? 20 Gbps available for a P2P service between Caltech and Fermilab Request 20 Gbps P2P service between Caltech and Fermilab. If 20 Gbps not available 10Gbps is ok. Failure on a network element, problem fixed 15 Gbps P2P service between Caltech and Fermilab Instantiated 20 Gbps available for a P2P service between Caltech and Fermilab 15 Gbps P2P service between Caltech and Fermilab Instantiated Endpoint Listing Service is not working, please check status Failure on a network element, problem fixed

SENSE Orchestrator and Services Application Workflow Agent Data Plane Connectivity Services: Point-to-Point Multi-Point Options Layer 2 Layer 3 Addressing (on top of Layer 2) Layer 3 Routed Network Connections Quality of Service Scheduling Preemption Strict and Loose hops Negotiation Batch Service Request Lifecycle monitoring and debug SENSE-O is based on StackV open source model driven orchestration system: github.com/MAX-UMD/stackV.community

SENSE Network Resource Manager (RM) SENSE-O Network-RM Functions/Roles: Responsible for a specific set of Network Resources Generate realtime MRML Model Evaluate and respond to SENSE-O information and service requests (including negotiation) Provision network resources in support of SENSE services Provide status, monitoring, and debug functions SENSE-RM API Model Based Interface Infrastructure and Services Network-RM OSCARS/NSI SDN Controller Others Network Controller

SENSE DTN/End Site Resource Manager SENSE-O EndSite/DTN-RM Functions/Roles: Responsible for a specific set of EndSite and DTN Resources Generate realtime MRML Model Evaluate and respond to SENSE-O information and service requests (including negotiation) Provision EndSite/DTN resources in support of SENSE services (includes networking stack of end systems) Provide status, monitoring, and debug functions QoS provided via OpenFlow (Open vSwitch) flow prioritization and/or TC (FireQoS) SENSE-RM API Model Based Interface Infrastructure and Services EndSite/DTN RM To external network Switch/Router End-Site/DTN Resources

Actual Demonstration Topology is Complex!

But not as complex as real systems Initial SENSE Use Cases LHC, CMS 13 Tiers 1s, 170 Tier 2s Superfacility

End to End Provisioning (see us for a demonstration) Workflow: -SENSE-RM provide Resource models to SENSE-O -SENSE-O builds end-to-end Model -Batch Provisioning of three P2P connections (UMD/MAX to Caltech-Booth, FNAL to Caltech-Booth, FNAL to Caltech) -Dataflow between end systems using FDT SENSE End-to-End Model SENSE Orchestrator RM DTN [200G] RM DTN [40G] RM DTN [40G] RM DTN [50G] UMD RM RM RM ESnet MAX DTN [200G] RM SENSE Resource Manager

SENSE Project Status Start of second year of three year project Initial versions of the following complete: Network Resource Manager, EndSite/DTN Resource Manager, Orchestrator Following functions complete: Resource Model Generation and processing, resource computation, abstract service requests, basic layer 2 P2P service instantiation To do list: negotiation, additional QoS, multi-point, resource computation for more complex services and questions, scheduling, lifecycle services monitoring and troubleshooting, application workflow agent and northbound API, preemption, strict/loose hops, application integration

Interesting Research Questions SENSE Interesting Research Questions Tradeoffs between scaling and real-time performance How to make this dynamic/configurable to adjust based on conditions and application needs Which information/states should be routinely exchanged between Resource Manager(s) and Orchestrator? Which information should be accessible on demand? How to make this dynamic/configurable to adjust based on different Resource Manager capabilities and policies? What is the right level of abstraction for Application Agent to SENSE system interactions? Is it necessary to provide a variable levels from highly abstract to very detailed and resource specific? What is the best method for realizing multi-domain, multi-resource authentication and authorization? What is the proper granularity for this? User? Project? Domain? Individual network and end system resource elements? Flows?

Thanks!