Presentation is loading. Please wait.

Presentation is loading. Please wait.

Steve Cotter, Chin Guok, Joe Metzger, Bill Johnston

Similar presentations


Presentation on theme: "Steve Cotter, Chin Guok, Joe Metzger, Bill Johnston"— Presentation transcript:

1 Steve Cotter, Chin Guok, Joe Metzger, Bill Johnston
ESnet Update Steve Cotter, Chin Guok, Joe Metzger, Bill Johnston

2 Agenda Network Update OSCARS perfSONAR Federated Trust

3 (USLHCnet: DOE+CERN funded)
ESnet4 – Jan 2009 SINet (Japan) Russia (BINP) CERN/LHCOPN (USLHCnet: DOE+CERN funded) GÉANT - France, Germany, Italy, UK, etc CA*net4 France GLORIAD (Russia, China) Korea (Kreonet2 MREN StarTap Taiwan (TANet2, ASCGNet) AMPATH CLARA (S. America) CUDI (S. America) Japan (SINet) Australia (AARNet) Canada (CA*net4 Taiwan (TANet2) Singaren Transpac2 CUDI KAREN/REANNZ ODN Japan Telecom America NLR-Packetnet Internet2 Korea (Kreonet2) KAREN / REANNZ Transpac2 Internet Korea (kreonet2) SINGAREN Japan (SINet) ODN Japan Telecom America CA*net4 GÉANT in Vienna (via USLHCNet circuit) LBL SLAC JLAB PPPL Ames ANL StarLight MAN LAN (32 A of A) PNNL BNL ORNL FNL LLNL LANL GA Yucca Bechtel-NV IARC INL NSTEC Pantex SNLA DOE-ALB Allied Signal KCP SRS NREL DOE NETL NNSA ARM ORAU OSTI NOAA Lab Link MAN NLR 10G 20G SDN SDN IP Peering Link IP router Lab Optical node SDN router

4 2008 Hub & Wave Install Timeline
JUL AUG SEP OCT NOV DEC STAR MX480 IP MX960 SDN 2008 HUB Installs 6 MX480’s 19 MX960’s SITE Installs 1 M120 PPPL 1 M10i LASV-HUB CHIC MX960 SDN DENV MX960 IP ATLA MX960 SDN NASH MX960 2nd set of Juniper MX’s arrived at LBNL in Mid Sept LASV PNWG MX960 6 10GE Internet2 waves installed, split & accepted 1 10G Framenet XC in WASH M10i WASH MX960 NEW HUB INSTALL 1st set of Juniper MX’s arrived at LBNL in Mid June AOFA MX480&MX960 EXISTING HUB UPGRADE NEWY MX960 BOST MX480 2 10GE Internet2 waves installed, split & accepted 1 10GE NLR AOFA-WASH 1 ORNL-NASH 10G IP 1 OC12 LANV-SUNN 1 10GE Internet2 STAR-CHICHIC CLEV MX960 x2 KANS MX480 MX960 HOUS 14 10GE Internet2 wave installed/split & accepted MX960 x2 PPPL ELPA MX960 x2 1 10GE MAN-LAN #2 1 10GE NRL Temp WASH-STAR 1 10GE CIC-OMNIPop at STAR ALBU MX480 M120 SDSC MX480 19 10GE Internet2 waves installed/split & accepted LOSA MX960 BOIS MX960 DENV MX960 ESnet Confidential Created by Mike O’Connor Mod by JimG

5 Hub & Wave Count Current Hub Count: Current Backbone Wave Count:
21 Completed: 32 AofA, NEWY*, WASH, ATLA, NASH, CLEV*, BOST*, CHIC, STAR, KANS*, HOUS*, ELPA*, DENV, ALBU, BOIS*, PNWG, SUNN, SNV(Qwest), LOSA*, SDSC, LASV(SwitchNap)* *9 New Hubs since July 2008 Current Backbone Wave Count: Internet2 / Level3 Waves: IP Waves: 17 new/split for a total of 25 SDN Waves: 25 new/split for a total of 30 NLR Waves: 1 new wave for a total of 5 1 temp wave (STAR-WASH) for used during NLR northern path upgrade

6 MAN Upgrades Timeline JAN FEB MAR LBNL LBNL SNV SNLL LLNL SLAC FNAL
MX960 LBNL MX960 LBNL 1 10GE LIMAN#3 AofA-BNL IP up Feb 2nd 1 DF circuit between AofA-NEWY up on Feb 2nd 1 LIMAN#4 NEWY-BNL (Waiting on Lightower to complete early Feb.) MX960 SNV MX480 SNLL MX480 LLNL MX480 SLAC MX960 FNAL LBL-MR2, SNV-MR1, SNLL-MR2, LLNL-MR2 & SLAC-MR2 (Completed on or before Jan 27th) MX480 FNAL MX960 NERSC MX480 JGI MX960 BNL FNAL’s MX Shipping Feb 3rd MX480 BNL MX960 ANL Final BAMAN 6509 replacements mid Feb. 1 10GE FRPG (upgrade from 1GE) DENV BNL & ANL install TBD ESnet Confidential Created by Mike O’Connor Mod by JimG

7 Active ESnet Links as of 12/2008
Link speed Description Count 10 GE National Core Waves (inter-hub) 61 Metropolitan Area Network Circuits (SF Bay Area MAN, Chicago MAN, Long Island MAN) 33 Circuits to ESnet sites 24 Circuits to R&E peering points Total 10G WAN circuits 125 Intra-hub connections (interconnecting ESnet equipment at the network hubs) 78 5 GE GÉANT peering in Vienna, Austria (via USLHCNet and GÉANT circuits) 1 OC-192 SONET special OC-48 SONET ORNL backup circuit 1 GE Mostly small site and commercial peering connections 83 Misc. slower Mostly non-SC sites 64

8 Future Installs Replace site 6509s (FNAL, ANL & BNL) with MXs
FNAL (MX960 & MX480) shipped on Feb 3rd for site to install BNL (MX960 & MX480) shipping & Install TBD ANL (MX960) shipping & Install TBD Replace BAMAN 6509s with MXs LBNL-MR3 (MX960), SNV-MR2 (MX960), LLNL-MR2 (MX480) & SNLL-MR2 (MX480) completed prior to Jan 22nd SLAC-MR2 (MX480) Completed on Jan 27th NERSC-MR2 & JGI-MR2 installs scheduled for Mid Feb. Future Circuits installs New 10 G LIMAN wave & DF AOFA-NEWY End-2-end on Feb 2nd & #4 wave to BNL (Feb) OC-12 between LASV hub and General Atomic (Feb) 10 GE between BOST hub to MIT (Feb) OC-12 between DENV hub and Pantex (TBD) 1 GE wave in BOIS to INL via IRON (TBD) 10 GE SDN wave between PNWG hub to PNNL (TBD) 10 GE SDN wave between NASH hub to ORNL (TBD) ESnet Confidential

9 San Francisco Bay Area MAN
ESnet4 Metro Area Rings Long Island MAN BNL 111- 8th (NEWY) 32 AoA, NYC USLHCNet USLHCNet FNAL 600 W. Chicago Starlight ANL West Chicago MAN LI MAN expansion, BNL diverse entry FNAL and BNL dual ESnet connection Upgraded Bay Area MAN switches San Francisco Bay Area MAN LBNL SLAC JGI LLNL SNLL NERSC SUNN SNV1 Newport News - Elite JLab ELITE ODU MATP Wash., DC Atlanta MAN ORNL (backup) Nashville 56 Marietta (SOX) Wash., DC 180 Peachtree Houston

10 Tier1 Redundancy: the Northeast
Boston / MIT To Seattle MAN LAN (A of A) BNL Clev. 111 8th, NYC To Chicago 32 A of A, NYC Wash. DC To Atlanta

11 Tier1 Redundancy: Long Island
Notes: There are physically independent paths from R1 to Boston and from R2 to Washington Only fiber paths are shown, wave counts are not indicated The engineering and procurement for this configuration are complete, implementation is underway An architecturally similarly situation is also being implemented for FNAL / Chicago To CERN Long Island To Boston 111 8th, NYC R1 USLHCNet R2 BNL IP core node SDN core node To Chicago ESnet IP core ESnet Science Data Network core (N X 10G) ESnet SDN core, NLR links (backup paths) Other R&E supplied link LHC related link MAN link International IP Connections 32 A of A, NYC To Washington

12 12 Month Circuit Availability 1/2009

13 Improved Site Availability

14 ESnet Accepted Traffic (Tby/mo)
Top 1000 Flows

15 Historical ESnet Traffic Patterns
Log Plot of ESnet Monthly Accepted Traffic, January 1990 – December 2008 Terabytes / month Oct 1993 1 TBy/mo. Aug 1990 100 MBy/mo. Jul 1998 10 TBy/mo. 38 months 57 months 40 months Nov 2001 100 TBy/mo. Apr 2006 1 PBy/mo. 53 months Traffic Increases 10X Every 47 Months July 2010 10 PBy/mo. 15

16 Network Traffic, Science Data, and Network Capacity
Projection 4 Pb/y 2010 Historical 40 Pb/y 2010 Climate model data All Four Data Series are Normalized to “1” at Jan. 1990 ESnet traffic HEP experiment data ESnet capacity roadmap (HEP data courtesy of Harvey Newman, Caltech, and Richard Mount, SLAC. Climate data courtesy Dean Williams, LLNL, and the Earth Systems Grid Development Team.)

17 ESnet4 – 2010 StarLight MAN LAN PNNL 40 40 BNL 40 50 50 (32 A of A) 40
FNL ORNL LLNL GA BNL LANL 40 40 40 50 50 40 50 50 50 50 30 40 40 30 50 40 30 40 30 40 40 Lab Link MAN NLR 10G 30/40/50G SDN IP IP router Lab Optical node SDN router 40 40

18 Beyond 2010: 100 G ESnet4 planning assumes technology advances will provide 100 Gb/s optical waves (they are 10 Gb/s now) The ESnet4 SDN switching/routing platform (Juniper MX960) is designed to support new 100 Gb/s network interfaces With capacity planning based on the ESnet 2010 wave count, we can probably assume some fraction of the core network capacity by 2012 will require 100 Gb/s interfaces ESnet is involved in a collaboration with Internet2, Juniper Networks (core routers), Infinera (DWDM), and Level3 (network support) to accelerate its deployment and help drive down the cost of 100G components

19 ESnet Security & Disaster Recovery
Advances in security at ESnet over the last 6 months: Implemented Two-factor authentication for ESnet network engineers requesting privileged access to the network management plane. Reviewed and re-defined access to network management plane. Upgraded Bro Intrusion Detection System ESnet Security Peer Review – Feb 11-12 Fed/R&E/Commercial experts reviewing ESnet security practices and procedures   Disaster recovery improvements Deployed Government Emergency Telecommunications Service (GETS) numbers to key personnel Deploying full replication of the NOC databases and servers and Science Services databases in the NYC Qwest carrier hub

20 Website Redesign Goals
Better organization of information, easier navigation, searchable (not everything in pdfs) but don’t want it to all be ‘push’ Collaborative tool – upload best practices, video from conference, community calendar, staff pages Integration of business processes into site “My ESnet” portal for site coordinators / users Exploring Google Earth or similar network visualization IP / SDN / MAN representation perfSONAR performance data OSCARS virtual circuit status Looking for ideas/input/suggestions.

21 Agenda Network Update OSCARS perfSONAR Federated Trust

22 Multi-Domain Virtual Circuit Service OSCARS
The OSCARS service requirements: Guaranteed bandwidth with resiliency User specified bandwidth - requested and managed in a Web Services framework Explicit backup paths can be requested Traffic isolation Allows for high-performance, non-standard transport mechanisms that cannot co-exist with commodity TCP-based transport Traffic engineering (for ESnet operations) Enables the engineering of explicit paths to meet specific requirements e.g. bypass congested links; using higher bandwidth, lower latency paths; etc. Secure connections The circuits are “secure” to the edges of the network (the site boundary) because they are managed by the control plane of the network which is highly secure and isolated from general traffic End-to-end, cross-domain connections between Labs and collaborating institutions

23 OSCARS Current (v0.5) Implementation
IDC InterDomain Controller User App User Well defined inter-module interfaces Exchange of static topology information PCE integrated into OSCARS Core ESnet Public WebServer (Proxy) Source Sink SDN IP IP Link SDN Link ESnet WAN Notification Call-back Event API WBUI Web Based User Interface Notification Broker API Resv API WS Interface OSCARS Core Reservation Management Path Computation Scheduling Inter-Domain Communications NS NotificationSubsystem PSS Path Setup Subsystem Network Element Interface AAAS Authentication Authorization Auditing Subsystem ESnet IDC (OSCARS) HTTPS HTTPS (SOAP) RMI SSHv2

24 OSCARS Future Implementation
IDC InterDomain Controller User App User Exchange of dynamic topology information includes time dimension PCE separated from OSCARS Core PCEs can be daisy changed allows PCE to be pluggable facilitates a research framework for collaboration ESnet Public WebServer (Proxy) Source Sink SDN IP IP Link SDN Link ESnet WAN Notification Call-back Event API WBUI Web Based User Interface Notification Broker API Resv API WS Interface OSCARS Core Reservation Management Scheduling Inter-Domain Communications NS NotificationSubsystem PSS Path Setup Subsystem Network Element Interface AAAS Authentication Authorization Auditing Subsystem PCE Path Computation Engine ESnet IDC (OSCARS) HTTPS HTTPS (SOAP) RMI SSHv2

25 Production OSCARS Modifications needed by FNAL and BNL
Changed the reservation workflow, added a notification callback system, and added some parameters to the OSCARS API to improve interoperability with automated provisioning agents such as LambdaStation, Terapaths and Phoebus. Operational VC support As of 12/2/08, there were 16 long-term production VCs instantiated, all of which support HEP 4 VCs terminate at BNL 2 VCs support LHC T0-T1 (primary and backup) 12 VCs terminate at FNAL For BNL and FNAL LHC T0-T1 VCs, except for the ESnet PE router at BNL (bnl-mr1.es.net) and FNAL (fnal-mr1-es.net), there are no other common nodes (router), ports (interfaces), or links between the primary and backup VC. Short-term dynamic VCs Between 1/1/08 and 12/2/08, there were roughly 2650 successful HEP centric VCs reservations 1950 reservations initiated by BNL using Terapaths 1700 reservations initiated by FNAL using LambdaStation

26 OSCARS is a Production Service
Site VLANS ESnet PE OSCARS setup all VLANs ESnet Core USLHCnet VLANS USLHCnet VLANS Tier2 LHC VLANS USLHCnet VLANS USLHCnet VLANS Tier2 LHC VLANS T2 LHC VLAN USLHCnet (LHC OPN) VLAN OSCARS generated and managed virtual circuits at FNAL – one of the US LHC Tier 1 data centers. This circuit map (minus the yellow callouts that explain the diagram) is automatically generated by an OSCARS tool and assists the connected sites with keeping track of what circuits exist and where they terminate.

27 Spectrum Now Monitors OSCARS Circuits

28 Agenda Network Update OSCARS perfSONAR Federated Trust

29 perfSONAR Services End-to-end monitoring service: providing useful, comprehensive, and meaningful information on the state of end-to-end paths. Supports regularly scheduled tests & archiving of results, acting as an intermediate layer, between the performance measurement tools and the diagnostic or visualization applications. Tools in the perfSONAR software suite:  SNMP Measurement Archive Lookup Service Topology Service Circuit Status Measurement Archive Status Measurement Archive perfSONAR-BUOY PingER Services Visualization Allow ESnet user community to better understand our network & its capabilities. Allow ESnet users to understand how their use impacts the backbone. Alarming Automated analysis of regularly scheduled measurements to raise alerts.

30 ESnet Deployment Activities
Currently deploying the hardware across the network to support adhoc measurements for debugging OWAMP Servers BWCTL Servers Topology Service Utilization Service perfSONAR Buoy Deployment Between ESnet systems To Internet2 & GEANT To/From ESnet Sites Hardens the infrastructure Continuous monitoring of servers & services Centralized management of OS & Services configuration Performance tuning & verifying everything is working as designed

31 perfSONAR R&D Activities
Scaling & robustness enhancements Visualization Tools Single Domain Tools Utilization Browser Topology Browser Latency & Bandwidth Browser Advanced Tools Looking across multiple domains Looking at correlations between different types of measurements Application or user community specific views Alarming Integrating OSCARS circuits Topology Utilization Active measurements across them

32 Federated Trust Services
Agenda Network Update OSCARS perfSONAR Federated Trust Services

33 Federated Trust Services
DOEGrids Certification Authority New Logo and ID Mark Operations DOEGrids Audit progress Cloning and Geographical Dispersion OpenID and Shibboleth Authorization Services Profile Document

34 DOEGrids CA - Operations
Vista – IE browser support in development Also beginning testing IE 8 browser ESnet 2-factor Support ESnet 2-factor authentication token project Add ESnet RA to list of official RAs in DOEGrids CA Recent problems – Dec 2008 CA not reading own Cert Revocation Lists CA automatically certifying customers from a peer, highly trusted CA (CERN CA) These problems have been corrected All certifications since June 2007 were audited No fraudulent certifications were discovered By agreement with registration authorities, affected subscribers will undergo direct reverification at next renewal (RA’s are free to require this at any time) (See auditing slide)

35 DOEGrids CA (one of several CAs) Usage Statistics
User Certificates 9259 Total No. of Revoked Certificates 2056 Host & Service Certificates 21043 Total No. of Expired Certificates 19452 Total No. of Requests 35629 Total No. of Certificates Issued 30331 Total No. of Active Certificates 8823 ESnet SSL Server CA Certificates 50 FusionGRID CA certificates 113 * Report as of Jan 29, 2009

36 DOEGrids CA (Active Certificates) Usage Statistics
* Report as of Jan 29, 2009

37 Active DOEGrids CA Breakdown
** OSG Includes (BNL, CDF, CIGI,CMS, CompBioGrid, DES, DOSAR, DZero, Engage, Fermilab, fMRI, GADU, geant4, GLOW, GPN, GRASE, GridEx, GUGrid, i2u2, ILC,  JLAB, LIGO, mariachi, MIS, nanoHUB, NWICG, NYSGrid, OSG, OSGEDU, SBGrid, SDSS, SLAC, STAR & USATLAS)

38 DOEGrids CA - Audits The Certification Practices Statement (CPS) is being “translated” to the RFC 3647 format Audit finding – requirement Appropriate format for interoperation Next step will be to correct all documentation errors identified in the audit Scheduling an audit of configurations, modules, and operational scripts (see Dec 2008 problems) – Feb/Mar 2009

39 DOEGrids CA Cloning & Geographical Dispersion
DOEGrids CA and its key management hardware will be cloned and dispersed around the US Improve Continuity of Operations and disaster recovery issues (ESnet requirements) Improve availability to customers Provision for future, robust services Current status: Testing and configuration of netHSM hardware, and project planning

40 OpenID and Shibboleth Continue efforts to promote this technology in DOE Laboratory community – won’t you join us? OpenID: Summer project testing OpenID provider (mostly) with simple server Objective: Use DOEGrids CA as source of identity Objective: Test simple application (custom, and later simplified wiki) See Roadmap for phase 2: Robust version of summer project, with more SSO and addition of other OpenID consumers as opportunities appear Shibboleth: Similar roadmap as for OpenID Many security issues to consider WAYF/Discovery a problem for both services – perhaps this is an opportunity for a 3rd service, CardSpace

41 Identity and Federation Technology
Shibboleth SAML 2.0 InCommon Federation OpenID OP and demo Consumer Graphics from SWITCH

42


Download ppt "Steve Cotter, Chin Guok, Joe Metzger, Bill Johnston"

Similar presentations


Ads by Google