Presentation is loading. Please wait.

Presentation is loading. Please wait.

The extension of optical networks into the campus Wade Hong Office of the Dean of Science Carleton University.

Similar presentations


Presentation on theme: "The extension of optical networks into the campus Wade Hong Office of the Dean of Science Carleton University."— Presentation transcript:

1 The extension of optical networks into the campus Wade Hong Office of the Dean of Science Carleton University

2 Outline Motivation CA*net 4 IGT From the Carleton U Perspective Issues Lessons learned

3 Motivation Large scale distributed scientific experiments (LHC - ATLAS, SNOLab, Polaris, NEES Grid... ) Access to regional distributed HPC resources (HPCVL, SharcNet, WestGrid, TRIUMF Tier 1.5,...) Federating growing research-based computing resources on campus Allowing the end users to access these resources in an unencumbered way CA*net 4 customer empowered networking last mile

4 CA*net 4 IGT CANARIE funded directed research project build a testbed to experiment with customer empowered networking, pt2pt optical networks, network performance, long haul 10 GbE, UCLP, last mile issues, etc. participants from the HEP community from across Canada, the provincial ORANs, CERN, StarLight, SURFnet, and potentially others setup end to end GbE and 10 GbE lightpaths between institutions in Canada and CERN

5 CA*net 4 Network

6 CA*net 4 IGT Sites

7 CA*net 4 IGT interoperability testing with 10 GbE WAN PHY and OC-192 used IXIA traffic generators to characterize the trams-atlantic link transferred real experimental data from ATLAS FCAL beam tests (GbE and 10 GbE) demonstrated native end-to-end 10 GbE between CERN and Ottawa for the ITU Telecom World 2003

8 Planned CA*net 4 IGT Activities complete the last mile connectivity for most of the participating Canadian sites third OC-192 across Canada being brought up using Nortel OME 6500s continuing long haul native 10 GbE experiments (Foundry MG8s) TRIUMF to CERN, TRIUMF to Carleton, Carleton to CERN CERN to Tokyo via Canada HEPix Robust Transfer Challenge - sustained disk to disk transfers between TRIUMF and CERN

9 Planned CA*net 4 IGT Activities Real-time remote farms for ATLAS CERN to U of Alberta Data transfer of End Cap Calorimeter data from the combined beam tests to several Canadian sites one beam test just completed (~1TB) second test to start late August (significantly more data) Transfer of CDF MC data from the Big Mac Cluster establish a GbE lightpath between UofT and FermiLab

10 Planned CA*net 4 IGT Activities Experimentation with bulk data transfer investigating RDMA/IP (sourcing NICs) establish GbE lightpaths between Canadian sites

11 Carleton University located in Ottawa, the nation’s capital at the southern end of the world’s longest outdoor skating rink Canada’s Capital University student population of 22,000 students, 1700 faculty and staff over $100M in research funding in the past year CFI contribution significant about half to Physics Bill St. Arnaud’s alma mater

12 Carleton University

13 External Network Connectivity commodity Internet Telecom Ottawa - was the largest metro 10 GbE deployment R&E traffic finally connected to ORION (Dec 2003), the new ORAN, just prior to the decommissioning of ONET EduNet non profit, OCRI managed dial-up and High Speed Internet for higher education institutions in Ottawa dial-up ISP has a dedicated link back to campus

14 Carleton U Network Upgrade campus has been in the process of planning a campus network upgrade for the past 3 to 4 years several false starts application to funding agencies based on requirements of research activities may have missed the window of opportunity finally proceeding with the network upgrade RFPs currently being evaluated

15 Network Upgrade Proposal original proposal phase one (Year 1) build the campus core network phase two (Year 2) build the distribution layer phase three (Year 3) rewire the buildings for access not my preferred ordering!

16 Proposed Topology

17 Differing Viewpoints debate over how to handle high capacity research traffic flows necessity of routing traffic through the proposed high capacity campus core on the other hand optical bypasses would simplify and reduce the complexity and cost of the campus network 4 fibre pairs between Herzberg Laboratories and Robertson Hall cost about $4K CDN - we prevailed reality check current campus network cannot handle the high volume and high speed flows

18 Motivations Revisited Large scale distributed scientific experiments

19 Motivations Revisited Access to regional distributed HPC resources other HPCVL sites (Queens, UofO, RMC, Ryerson U) TRIUMF ATLAS Canada computing centre SNOLab shared ORION and CA*net 4 connectivity is only at GbE high capacity flows probably dictate pt2pt optical bypass interconnectivity can be static or dynamic fully statically meshed or scheduled dynamic connectivity on demand - probably the latter

20 Motivations Revisited Federating growing research-based computing resources into a campus grid HPCVL Linux cluster upgrade (128+256 CPUs) Physics research cluster upgrade (40+96+96 CPUs) Civil Engineering (~128 CPUs) Architecture/Psychology visualization cluster (>128 CPUs) Systems and Computer Engineering ( 64 CPUs) debating a condominium or distributed model most likely a hybrid with optical fibre as the interconnecting fabric probably static pt2pt optical bypass for ease of use and user control

21 Motivations Revisited federated the Physics research computing cluster with part of the HPCVL Linux cluster last summer for about 2 months clusters located on different floors pt2pt link established - much easier than routing through the campus network completed half of the MC regeneration for the third SNO paper similar arrangement this summer to add part of the HPCVL cluster to the Carleton U Physics contribution to the LHC Computing Grid till the end of the year

22 Issues control central management and control vs end user empowerment disruptive network complexity using pt2pt ethernet links for high capacity flows should simplify campus networks (reduce costs?) security disruptive - bypassing DMZ for the uses considered here, the pt2pt links are inherently secure - non routed private subnets

23 Issues why not copper? it could be but with fibre greater distances requires less active devices along the path management and control - device at each end under the control of the end users is ideal consistent device characteristics - jumbo frames, port speed, duplex, etc. inter-building connectivity is fibre and planned vertical cabling will be fibre

24 Issues last mile connectivity demarcation point end user device (NIC) or an edge device (switch, CWDM mux) location of the demarc at the end user or a common shared location technology used to extend the end to end lightpath into the campus pt2pt GbE optical GbE NIC - patched thru to GbE interface on ONS media converter - copper to optical

25 Issues pt2pt 10GbE LAN PHY to WAN PHY conversion to OC192c on ONS 15454/OME 6500 wavelength conversion CWDM media converters - copper to colored wavelength colored GBICS for GbE switch optical link charateristics padding (attenuation), proper power budget, etc. end user shouldn’t need to be an optical networking expert

26 Lessons Learned good to be rich in fibre provides greater flexibility support of ORANs, national R&E network, and international partners is essential - all have been very supportive need to convince local campus networking folks that this is not really too disruptive will simplify and not burden the campus production network need a more coherent way of dealing with optical access in the last mile still lots to learn!

27 Thank You! Wade Hong xiong@physics.carleton.ca


Download ppt "The extension of optical networks into the campus Wade Hong Office of the Dean of Science Carleton University."

Similar presentations


Ads by Google