LAN QoS and WAN MPLS: Status and Plan Dantong Yu and Shawn Mckee DOE Site Visit December 13, 2004.

Slides:



Advertisements
Similar presentations
LambdaStation Phil DeMar Don Petravick NeSC Oct. 7, 2004.
Advertisements

Bernd Panzer-Steindel, CERN/IT WAN RAW/ESD Data Distribution for LHC.
Authenticated QoS Signaling William A. (Andy) Adamson Olga Kornievskaia CITI, University of Michigan.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
TeraPaths TeraPaths: Flow-Based End-to-End QoS Paths through Modern Hybrid WANs Presented by Presented by Dimitrios Katramatos, BNL Dimitrios Katramatos,
QoS Solutions Confidential 2010 NetQuality Analyzer and QPerf.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Chin Guok ESnet Network Engineer David Robertson DSD Computer Software Engineer Lawrence Berkeley National Laboratory.
TeraPaths: End-to-End Network Path QoS Configuration Using Cross-Domain Reservation Negotiation Bruce Gibbard Dimitrios Katramatos Shawn McKee Dantong.
1 Chin Guok ESnet Network Engineer David Robertson DSD Computer Software Engineer Lawrence Berkeley National Laboratory.
GNEW 2004 CERN, Geneva, Switzerland March 16th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview for GNEW2004 Shawn McKee University of Michigan.
Technology Steering Group January 31, 2007 Academic Affairs Technology Steering Group February 13, 2008.
Hands-On Microsoft Windows Server 2003 Networking Chapter 7 Windows Internet Naming Service.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 11 Managing and Monitoring a Windows Server 2008 Network.
TeraPaths : A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research USATLAS Tier 1 & Tier 2 Network Planning Meeting December.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Circuit Services - IPTV Christian Todorov Internet2 Fall Member Meeting October 9, 2007.
©Kwan Sai Kit, All Rights Reserved Windows Small Business Server 2003 Features.
TeraPaths: A QoS Collaborative Data Sharing Infrastructure for Petascale Computing Research Bruce Gibbard & Dantong Yu High-Performance Network Research.
© 2006 Cisco Systems, Inc. All rights reserved.Cisco Public 1 Version 4.0 Identifying Application Impacts on Network Design Designing and Supporting Computer.
TeraPaths TeraPaths: establishing end-to-end QoS paths - the user perspective Presented by Presented by Dimitrios Katramatos, BNL Dimitrios Katramatos,
Lambda Station Project Andrey Bobyshev; Phil DeMar; Matt Crawford ESCC/Internet2 Winter 2008 Joint Techs January 22; Honolulu, HI
Apr 30, 20081/11 VO Services Project – Stakeholders’ Meeting Gabriele Garzoglio VO Services Project Stakeholders’ Meeting Apr 30, 2008 Gabriele Garzoglio.
© 2006 Cisco Systems, Inc. All rights reserved.Cisco PublicITE I Chapter 6 1 Identifying Application Impacts on Network Design Designing and Supporting.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
USATLAS Network/Storage and Load Testing Jay Packard Dantong Yu Brookhaven National Lab.
Thoughts on Future LHCOPN Some ideas Artur Barczyk, Vancouver, 31/08/09.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
ASCR/ESnet Network Requirements an Internet2 Perspective 2009 ASCR/ESnet Network Requirements Workshop April 15/16, 2009 Richard Carlson -- Internet2.
TeraPaths TeraPaths: Establishing End-to-End QoS Paths through L2 and L3 WAN Connections Presented by Presented by Dimitrios Katramatos, BNL Dimitrios.
DYNES Storage Infrastructure Artur Barczyk California Institute of Technology LHCOPN Meeting Geneva, October 07, 2010.
Office of Science U.S. Department of Energy ESCC Meeting July 21-23, 2004 Network Research Program Update Thomas D. Ndousse Program Manager Mathematical,
1 Network Measurement Summary ESCC, Feb Joe Metzger ESnet Engineering Group Lawrence Berkeley National Laboratory.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
Practical Distributed Authorization for GARA Andy Adamson and Olga Kornievskaia Center for Information Technology Integration University of Michigan, USA.
BNL Tier 1 Service Planning & Monitoring Bruce G. Gibbard GDB 5-6 August 2006.
TeraPaths The TeraPaths Collaboration Presented by Presented by Dimitrios Katramatos, BNL Dimitrios Katramatos, BNL.
Connect. Communicate. Collaborate perfSONAR MDM Service for LHC OPN Loukik Kudarimoti DANTE.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
Online-Offsite Connectivity Experiments Catalin Meirosu *, Richard Hughes-Jones ** * CERN and Politehnica University of Bucuresti ** University of Manchester.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Terapaths: MPLS based Data Sharing Infrastructure for Peta Scale LHC Computing Bruce Gibbard and Dantong Yu USATLAS Computing Facility DOE Network Research.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team CHEP 06.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
1 TeraPaths and dynamic circuits  Strong interest to expand testbed to sites connected to Internet2 (especially US ATLAS T2 sites)  Plans started in.
Internet2 Joint Techs Workshop, Feb 15, 2005, Salt Lake City, Utah ESnet On-Demand Secure Circuits and Advance Reservation System (OSCARS) Chin Guok
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
-1- ESnet On-Demand Secure Circuits and Advance Reservation System (OSCARS) David Robertson Internet2 Joint Techs Workshop July 18,
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
HENP SIG Austin, TX September 27th, 2004Shawn McKee The UltraLight Program UltraLight: An Overview and Update Shawn McKee University of Michigan.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Grid Deployment Technical Working Groups: Middleware selection AAA,security Resource scheduling Operations User Support GDB Grid Deployment Resource planning,
Javier Orellana EGEE-JRA4 Coordinator CERN March 2004 EGEE is proposed as a project funded by the European Union under contract IST Network.
TeraPaths: A QoS Enabled Collaborative Data Sharing Infrastructure for Petascale Computing Research The TeraPaths Project Team Usatlas Tier 2 workshop.
Fermilab Cal Tech Lambda Station High-Performance Network Research PI Meeting BNL Phil DeMar September 29, 2005.
The TeraPaths Testbed: Exploring End-to-End Network QoS Dimitrios Katramatos, Dantong Yu, Bruce Gibbard, Shawn McKee TridentCom 2007 Presented by D.Katramatos,
Bob Jones EGEE Technical Director
“A Data Movement Service for the LHC”
Establishing End-to-End Guaranteed Bandwidth Network Paths Across Multiple Administrative Domains The DOE-funded TeraPaths project at Brookhaven National.
Unit 27: Network Operating Systems
The UltraLight Program
Presentation transcript:

LAN QoS and WAN MPLS: Status and Plan Dantong Yu and Shawn Mckee DOE Site Visit December 13, 2004

2 Outline  Introduction: The MPLS & LAN QoS project and personnel Involved.  BNL network configuration.  Program goals and details.  Proposed primitive infrastructure.  Work plans, milestones, schedule.  Current status.

3  This project will investigate the integration and use of LAN QoS and MPLS based differentiated network services in the ATLAS data intensive distributed computing environment as a way to manage the network as a critical resource;.  The Collaboration includes:  Brookhaven National Laboratory (US ATLAS Tier 1, ESNet)  USATLAS Group: Bruce Gibbard, Razvan Popescu, Tom Throwe, Dantong Yu, Xin Zhao  ITD Networking Group: Scott Bradley, Frank Burstein, Vincent Bonafede  Univ. of Michigan [UM] (US ATLAS Network Project Manager, US ATLAS Candidate Tier 2 Center, Internet2, UltraLight Co-PI): Shawn Mckee  The External Collaborators from OSCAR (ESNET), Lambda station (FNAL), and DWMI (SLAC) What is Terapaths ?

4 BNL Campus Network and Immediate Update Plan  The USATLAS/RHIC Computing Facility is attached to access/distribution layer routers.  Campus network consists of Cisco 6509 Series Switches, with Cisco QoS support  Cisco PIX 535 Firewall. (1Gbps)  Will be replaced by firewall service blade for CISCO 6500 series. (5 Gbps), Dec/14/2004

5 Project Goal and Objectives T The primary goal of this project is to investigate the use of QoS/MPLS technology in the ATLAS data intensive distributed computing environment. In addition we intend to: q Develop expertise in QoS technology and Management which will be important to ATLAS and the LHC community more generally. q Dedicate fractions of the available bandwidth via QoS supported by existing network infrastructure to ATLAS Tier 1 data movement and RHIC data replications to assure adequate throughput and limit their disruptive impact upon each other. q Enhance technical contact between the ATLAS tier 1 at BNL and its network partners including the Tier 0 center at CERN, potential ATLAS Tier 2’s and other members of the Grid3+ (OSG-0) community of which it is a part.

6 Proposed Prototype/Primitive Infrastructure GridFtp & SRM LAN/MPLS ESnet Network resource manager MPLS requests Traffic Identification TCP syn/fin packages, addresses, port # Grid AA Network Usage Policy Translator Bandwidth Requests & Releases OSCARS IN(E)GRESS Monitoring Direct Bandwidth Requests SE Second/Third year LAN QoS M10 LAN QoS requests

7 Overall Work Plans  Terapaths envisions a multiple year program to deliver a high- performance, QoS enabled network infrastructure for ATLAS/LHC computing. Each year will determine the following year(s)’s direction.  Phase I: Establish Initial Functionality (08/04 ~ 09/30). z Help to steer the direction of the following two phases.  Phase II:Establish Prototype Production Service (10/05 ~ 07/06).  Depends on the success of Phase 1.  Phase III:Establish Full Production Service, Extend Scope and Increase Functionality (08/06 ~ 07/07). z The level of service and its scope will depends on the available project funding and some additional resources. z Broaden deployment and capability to Tier2s, partners.

8 Goal: Have LAN QoS up and running with manual configuration, integrate with WAN’s MPLS PATH to provide End-to-end QoS (Oct/01~March/31: 6 months)  Milestone 1 and work items: (Jan/10/2005) Enable LAN QoS inside BNL campus network along the path from GridFtp servers and SRM into the boarder routers.  Identify QoS capabilities associated with LAN infrastructure. (ITD)  Study hardware-dependent QoS techniques, options for LAN QoS. (ITD, UM)  Initial deployment on testbed (non-production) routers to prove the concepts. (ITD,UM)  Configure local production network to enable QoS for pre-selected traffics. (ITD)  Milestone 2 and work items: (Jan/31/2005) Test and verify MPLS paths between BNL and LBL, SLAC (network monitoring project), FNAL.  Meet regularly with OSCAR, remote collaborators to request MPLS path setup. (BNL ATLAS, UM, OSCAR)  Send Traffic to MPLS paths provided by OSCAR. (BNL teams) Test and verify MPLS path between BNL and UM (BNL, OSCAR, AND UM). (March/30/2004)

9 6 months goal, milestones and work items  Milestone 3 and work items: (Jan/31/2005) Set up network monitoring tools provided by DWMI, install other monitoring tools when necessary.  Work with Les to attain and deploy the PingER and DWMI tools. (BNL and SLAC)  Install host monitoring tools, such as ganglia and network equipment monitoring tools. (BNL ATLAS and UM)  Milestone 4 and work items: (Jan/31/2005) Integrate the LAN QoS with MPLS paths offered by OSCAR.  Verify LAN QoS capability by injecting regular traffic and prioritized traffic in LAN. (BNL ATLAS and ITD)  Verify WAN MPLS bandwidth provisioning (BNL ATLAS and UM)  Do integration test and document experience. (BNL ATLAS and UM)  Create a simple LAN QoS and MPLS request system based on exchanges: list, conference. (BNL ATLAS and OSCAR)

10 6 months goal, milestones and work items  Milestone 5 and work items: (Feb/28/2005) Verify the impact between prioritized traffic and best effort traffic. Learn the effectiveness and efficiency of MPLS/LAN QoS and its impact to overall network performance  Create best effort traffic between BNL and CERN to fill up bottleneck bandwidth for prescribed periods. (BNL ATLAS)  Inject prioritized traffic between BNL and FNAL and monitor the traffic behavior. (BNL ATLAS and ITD)  Test and verify different LAN QoS technologies/policies in different time periods. (BNL ATLAS, ITD and UM)  Milestone 6 and work items: (March/31/2005) Add more applications into MPLS/LAN QoS projects  Add part of USATLAS/RHIC production data transfer into LAN/MPLS paths. (BNL ATLAS, ITD, OSCAR, RHIC)

11 Goal: Automate the MPLS/LAN QoS setup and raise aware of managing end-to-end network resource (April/01~Sep/31/05, 12 months)  Milestone 1: Document any lesson/experience learned from this project. (April/30/2005)  Decide whether LAN QoS/MPLS is useful to LHC physics. (BNL ATLAS, UM)  Build LAN QoS/MPLS expertise in BNL: LAN QoS & MPLS setup, configuration, maintenance and removal. (BNL ATLAS, ITD, UM)  Milestone 2: Implement automatic/manual QoS request system. (May/31/2005)  Traffic Identification: capture TCP syn/fin packages and register to the pending request queue. (BNL ATLAS, UM).  Dynamic generation of QoS requests in the midst of network traffic. (BNL ATLAS, SLAC)  Create Web interface for manually inputting QoS requests. (BNL ATLAS).  Interface into QoS scheduling System. (BNL ATLAS, UM).

12 12 Months goal, milestones and work items  Milestone 3: Implement simple bandwidth scheduling system. (June/30/2005)  Implement first come, first serve for LAN bandwidth. (BNL ATLAS)  Implement interface into OSCAR MPLS scheduler. (BNL ATLAS, ESNet)  Information collector for DWMI. (BNL ATLAS and SLAC)  Milestone 4: Implement automatic LAN QoS configuration systems. (June/30/2005)  Evaluate/Deploy vendor provided LAN management system software (BNL ITD, UM)  Map LAN QoS requests onto network control commands (BNL ITD, UM)

13 12 Months goal, milestones and work items  Milestone 5: System Integration, test, validation and deployment. (August/31/2005)  System integration onto available testbed equipment. (BNL, UM)  Test and Verify with synthetic traffic. (BNL)  Deploy the system into BNL production equipment (BNL, UM).  Milestone 6: Design a remote surveillance video demo to monitor BNL and FNAL SC2005 booths and home institutes and utilize WAN MPLS and LAN QoS. (BNL, FNAL) (Sep/31/2005)  Configure LAN QoS and MPLS paths between BNL and FNAL. (BNL, FNAL, Esnet)  Exchange video and audio streams between these two sites. (BNL, FNAL)  Generate physics event histogram from data stored at remote peers. (BNL, FNAL)  Compare with NON-QoS scenario and document the difference.

14 Establish Prototype Production Service (year 2) T Integrate Grid data transfer (GridFtp) into LAN QoS/MPLS enabled network service. T Effectively couple these network/data transfer services with storage resources managed by SRM, have GridFtp/SRM functionality in “beta”. T Incorporate the resulting system into the ATLAS grid middleware. T Build tools to provide basic authentication, authorization and access control – Depends on funding, Rely on leveraging. T Supply client interfaces which make this service available in a manner transparent to any details of the underlying QoS/MPLS traffic engineering T Leverage MPLS paths/VO level network monitoring services with DWMI project to be developed at SLAC.

15 Extend Scope and Increase Functionality of Prototype Service (year 3) T Inter-network domain MPLS establishment, dynamically creating, adjusting of sub-partitioning MPLS paths to meet time constrained network requirements. q Create site level network resource manager for multiple VOs vying for limited WAN resource. q Provide dynamic bandwidth re-adjusting based resource usage policy and path utilization status collected from network monitoring (DWMI). q Leverage dynamic MPLS establishing services provided by OSCARS: the ESnet On-Demand Secure Circuits and Advance Reservation System. q Create user interface/web services for LHC data transfer applications to request network resource in advance.  Goal: to broaden deployment and capability to tier 1 and tier 2 sites, create services which will be honored/adopted by CERN ATLAS/LHC Tier 0.

16 Status (Testbed and Labeling)  A LAN QoS simulation testbed is being created in a private network environment: consists of four hosts and two CISCO switches (same models as production infrastructure)  Evaluated mechanisms to assign different “cos” to GridFtp data transfers initiated by different VOs.  LAN: source IP addresses and port numbers determine DSCP bits  WAN MPLS: Use IP addresses, port # & DSCP bits for assign MPLS labels at border router. TEnable QoS: “mls qos” TClassify the packets at the receiving ports, 3/1 and 3/9. q Interface fa3/1: Mls qos cos 5 q Interface fa3/9: Mls qos cos 0 TBandwidth Policing: Fair queue tos 5 60 Fair queue tos 0 Class-map match-any green Match ip precedence 0 Class-map match-any red Match ip precedence 5 60% 40% 3/1 3/9

17 Status (Production network configuration)  LAN  packets from ftp servers (aftpexp01~04) and port (21,000~22,000) to FNAL CMS subnet were marked to class 4 and EF at three Cisco switches/routers.  Working on bandwidth allocation/policing for the marked traffic.  WAN  A MPLS path was manually created between the BNL ESNet router to Esnet POP to STARLight. (November/18/2004)  Traffic with correct source, destination ip addresses is correctly labeled and tunneled.  Bandwidth allocation is 50mbps, but policer does not correctly police the traffic.

18 Status (MPLS traffic verification)  Qualified pair of hosts: ~]$ /usr/sbin/traceroute cmssrv04.fnal.gov traceroute to cmssrv04.fnal.gov 1 anubis.s80.bnl.gov ( ) 2 shu.v400.bnl.gov ( ) 3 amon ( ) 4 esbnl-bnl.es.net ( ) 5 aoacr1-oc48-bnl.es.net ( ) MPLS Label= CoS=5 TTL=1 S=0 6 chicr1-oc192-aoacr1.es.net ( ) MPLS Label= CoS=5 TTL=1 S=0 7 chirt1-ge0-chicr1.es.net ( ) ( ) 10 10cmssrv04.fnal.gov ( )  Un-qualified pair of hosts: [ ~]$ /usr/sbin/traceroute cmssrv04.fnal.gov traceroute to cmssrv04.fnal.gov ( ), 30 hops max, 38 byte packets 1 anubis.s80.bnl.gov ( ) 2 shu.v400.bnl.gov ( ) 3 amon.bnl.gov ( ) 4 esbnl-bnl.es.net ( ) 5 aoacr1-oc48-bnl.es.net ( ) 6 chicr1-oc192-aoacr1.es.net ( ) 7 fnal-pos-chi.es.net ( ) 8 ge2-1.r-s-bdr.fnal.gov ( ) 9 vlan313.r-s-hub-fcc.fnal.gov ( ) 10 vlan312.r-cms-fcc1.fnal.gov ( ) 11 cmssrv04.fnal.gov ( ) ~]$ iperf -c cmssrv16.fnal.gov -p w 1MB -P 40 -t 100 -i 5 [SUM] sec 3.08 GBytes 259 Mbits/sec (much higher than allocated the 50mbps bandwidth)

19 Status ( Create best effort traffic between BNL and CERN to fill up bottleneck bandwidth for prescribed periods)  BNL participate into CERN Tier 0, tier 1 service challenge.  Prototype the data movement services that will be needed by LHC  Meet the required performance and reliability.  Network,  Disk to disk file transfer.  Reliable file transfer service.  Mass storage to mass storage file transfer.  Accomplished  Deploy and tune the four Ftp servers on INTEL based commodity PCs with SCSI drives, tune the kernel, local disk file systems.  Disk to disk file transfer between BNL and CERN GridFtp servers with GridFtp tools.  75MB/second writing to BNL disks and 80MB/second reading from BNL disk per host  Two servers filled up the network bottleneck 99%, went beyond OC 12 connection.  Sustained the data transfer rate for more than 12 hours. Stopped because of potential for disrupting other users.  BNL’s service challenge is scheduled during Christmas week.

20 Network Traffic Monitoring for “Robust Service Challenge”

21 Status (Network Monitoring Tools Deployment)  Deployed ganglia based host monitoring tools.  Requested ITD to evaluate Cisco proprietary network managing tools: such as COPS.  Have been working with SLAC DWMI team to deploy network monitoring tools. q IEPM_BW software is being installed on a machine outside the firewall with 100MBits interface, 50% installation was completed. SLAC team provides helps with the system installation. (Xin Zhao, Connie Logg)