Data Center Fabrics. Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use.

Slides:



Advertisements
Similar presentations
Data Center Fabrics Lecture 12 Aditya Akella.
Advertisements

PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric
Mobility Jennifer Rexford COS 461: Computer Networks Lectures: MW 10-10:50am in Architecture N101
J. K. Kim Portland. Trend of Data Center By J. Nicholas Hoover, InformationWeek June 17, :00 AMJ. Nicholas Hoover 200 million Euro Data centers.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. Presented by: Vinuthna Nalluri Shiva Srivastava.
Data Center Networking Major Theme: What are new networking issues posed by large-scale data centers? Network Architecture? Topology design? Addressing?
Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis Miri, Sivasankar Radhakrishnan, Vikram Subramanya, and Amin Vahdat Department.
PortLand: A Scalable Fault- Tolerant Layer 2 Data Center Network Fabric B 財金三 婁瀚升 1.
Applying NOX to the Datacenter Arsalan Tavakoli, Martin Casado, Teemu Koponen, and Scott Shenker 10/22/2009Hot Topics in Networks Workshop 2009.
Switching Topic 4 Inter-VLAN routing. Agenda Routing process Routing VLANs – Traditional model – Router-on-a-stick – Multilayer switches EtherChannel.
Multi-Layer Switching Layers 1, 2, and 3. Cisco Hierarchical Model Access Layer –Workgroup –Access layer aggregation and L3/L4 services Distribution Layer.
Network Overlay Framework Draft-lasserre-nvo3-framework-01.
Revisiting Ethernet: Plug-and-play made scalable and efficient Changhoon Kim and Jennifer Rexford Princeton University.
Datacenter Network Topologies
Virtual Layer 2: A Scalable and Flexible Data-Center Network Work with Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Parantap Lahiri,
Oct 21, 2004CS573: Network Protocols and Standards1 IP: Addressing, ARP, Routing Network Protocols and Standards Autumn
1 K. Salah Module 4.0: Network Components Repeater Hub NIC Bridges Switches Routers VLANs.
Portland: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric Offense Kai Chen Shih-Chi Chen.
CS335 Networking & Network Administration Tuesday, April 20, 2010.
Spring Routing & Switching Umar Kalim Dept. of Communication Systems Engineering 06/04/2007.
PortLand Presented by Muhammad Sadeeq and Ling Su.
Data Center Network Topologies: FatTree
ProActive Routing In Scalable Data Centers with PARIS Joint work with Dushyant Arora + and Jennifer Rexford* + Arista Networks *Princeton University Theophilus.
COS 461: Computer Networks
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
A Modernize Version of the Antique Telephone Architecture Christopher Francis & Bertha Wilhelm.
1 K. Salah Module 4.3: Repeaters, Bridges, & Switches Repeater Hub NIC Bridges Switches VLANs GbE.
A Scalable, Commodity Data Center Network Architecture Mohammad AI-Fares, Alexander Loukissas, Amin Vahdat Presented by Ye Tao Feb 6 th 2013.
A Scalable, Commodity Data Center Network Architecture
Jennifer Rexford Fall 2010 (TTh 1:30-2:50 in COS 302) COS 561: Advanced Computer Networks Data.
A Scalable, Commodity Data Center Network Architecture.
Datacenter Networks Mike Freedman COS 461: Computer Networks
Secure Cloud Computing with Virtualized Network Infrastructure HotCloud 10 By Xuanran Zong.
MPLS And The Data Center Adrian Farrel Old Dog Consulting / Juniper Networks
Networking the Cloud Presenter: b 電機三 姜慧如.
VL2 – A Scalable & Flexible Data Center Network Authors: Greenberg et al Presenter: Syed M Irteza – LUMS CS678: 2 April 2013.
© 2007 Cisco Systems, Inc. All rights reserved.Cisco Public ITE PC v4.0 Chapter 1 1 Connecting to the Network Networking for Home and Small Businesses.
Routing & Architecture
A Scalable, Commodity Data Center Network Architecture Jingyang Zhu.
LAN Switching and Wireless – Chapter 1
Floodless in SEATTLE : A Scalable Ethernet ArchiTecTure for Large Enterprises. Changhoon Kim, Matthew Caesar and Jenifer Rexford. Princeton University.
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
© 2008 Cisco Systems, Inc. All rights reserved.Cisco ConfidentialPresentation_ID 1 Chapter 1: Introduction to Scaling Networks Scaling Networks.
Department of Computer Science A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares Alexander Loukissas Amin Vahdat SIGCOMM’08 Reporter:
Authors: Xiaoqiao Meng, Vasileio Pappas and Li Zhang
Cisco Confidential © 2013 Cisco and/or its affiliates. All rights reserved. 1 Cisco Networking Training (CCENT/CCT/CCNA R&S) Rick Rowe Ron Giannetti.
[Fat- tree]: A Scalable, Commodity Data Center Network Architecture
Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis Miri, Sivasankar Radhakrishnan, Vikram Subramanya, Amin Vahdat SIGCOMM.
15-744: Computer Networking L-12 Data Center Networking I.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis.
Network Virtualization Ben Pfaff Nicira Networks, Inc.
Data Center Networking Major Theme: What are new networking issues posed by large-scale data centers? Network Architecture? Topology design? Addressing?
VL2: A Scalable and Flexible Data Center Network
Data Center Architectures
Data Center Networking
IP: Addressing, ARP, Routing
CIS 700-5: The Design and Implementation of Cloud Networks
Lecture 2: Cloud Computing
Data Center Network Topologies II
Lecture 2: Leaf-Spine and PortLand Networks
ETHANE: TAKING CONTROL OF THE ENTERPRISE
Revisiting Ethernet: Plug-and-play made scalable and efficient
NTHU CS5421 Cloud Computing
A Scalable, Commodity Data Center Network Architecture
湖南大学-信息科学与工程学院-计算机与科学系
NTHU CS5421 Cloud Computing
VL2: A Scalable and Flexible Data Center Network
Data Center Architectures
Reconciling Zero-conf with Efficiency in Enterprises
Data Center Traffic Engineering
Presentation transcript:

Data Center Fabrics

Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use standard intra-domain routing protocols, eg. OSPF. – Large administration overhead Layer 2 approach: Forwarding on flat MAC addresses Less administrative overhead Bad scalability Low performance – Middle ground between layer 2 and layer 3: VLAN Feasible for smaller scale topologies Resource partition problem

Requirements due to Virtualization End host virtualization: – Needs to support large addresses and VM migrations – In layer 3 fabric, migrating the VM to a different switch changes VM’s IP address – In layer 2 fabric, migrating VM incurs scaling ARP and performing routing/forwarding on millions of flat MAC addresses.

Motivation Eliminate Over-subscription – Solution: Commodity switch hardware Virtual Machine Migration – Solution: Split IP address from location. Failure avoidance – Solution: Fast scalable routing

Architectural Similarities Both approaches use indirection – Application address doesn’t change when VM moves, all that changes in Location address – Location addresses: specifies location in network – Application address: specifies address of VM A network of commodity switches – Reduces energy consumptions – Allows to afford enough switches to eliminate overprovision Central entity to perform name resolution between Location address and application address – Directory Service: VL2 – Fabric Manager: Portland – Both entities are triggered by ARP request. – Stores mapping of LA to AA Gateway devices – Perform encapsulation/decapsulation of external traffic

Architecture Differences Routing – VL2: Source routing based Each packet contains the address of all switches to traverse – Portland: topology based routing Location addresses encoding location with the tree Each switch is aware of how to decode location addresses – Forwarding is based on this intimate knowledge. Indirection – VL2: Indirection is on L3: IP-in-IP encapsulation – Portland: Indirection is on L2: IP-to-Pmac ARP functionality: – Portland: ARP returns IP to Pmac – VL2: ARP returns a list of intermediate switches to traverse

Portland

Fat-Tree Inter-connect racks (of servers) using a fat-tree topology Fat-Tree: a special type of Clos Networks (after C. Clos) K-ary fat tree: three-layer topology (edge, aggregation and core) – each pod consists of (k/2) 2 servers & 2 layers of k/2 k-port switches – each edge switch connects to k/2 servers & k/2 aggr. switches – each aggr. switch connects to k/2 edge & k/2 core switches – (k/2) 2 core switches: each connects to k pods Fat-tree with K=2 8

Why? Why Fat-Tree? – Fat tree has identical bandwidth at any bisections – Each layer has the same aggregated bandwidth Can be built using cheap devices with uniform capacity – Each port supports same speed as end host – All devices can transmit at line speed if packets are distributed uniform along available paths Great scalability: k-port switch supports k 3 /4 servers Fat tree network with K = 3 supporting 54 hosts 9

PortLand Assuming: a Fat-tree network topology for DC Introduce “pseudo MAC addresses” to balance the pros and cons of flat- vs. topology-dependent addressing PMACs are “topology-dependent,” hierarchical addresses – But used only as “host locators,” not “host identities” – IP addresses used as “host identities” (for compatibility w/ apps) Pros: small switch state & Seamless VM migration Pros: “eliminate” flooding in both data & control planes But requires a IP-to-PMAC mapping and name resolution – a location directory service And location discovery protocol & fabric manager – for support of “plug-&-play” 10

PMAC Addressing Scheme PMAC (48 bits): pod.position.port.vmid – Pod: 16 bits; position and port (8 bits); vmid: 16 bits Assign only to servers (end-hosts) – by switches 11 pod position

Location Discovery Protocol Location Discovery Messages (LDMs) exchanged between neighboring switches Switches self-discover location on boot up Location Characteristics Technique Tree-level (edge, aggr., core) auto-discovery via neighbor connectivity Position # aggregation switch help edge switches decide Pod # request (by pos. 0 switch only) to fabric manager 12

PortLand: Name Resolution Edge switch listens to end hosts, and discover new source MACs Installs mappings, and informs fabric manager 13

PortLand: Name Resolution … Edge switch intercepts ARP messages from end hosts send request to fabric manager, which replies with PMAC 14

PortLand: Fabric Manager fabric manager: logically centralized, multi-homed server maintains topology and mappings in “soft state” 15

VL2

Design: Clos Network Same capacity at each layer – No oversubscription Many paths available – Low sensitivity to failures

Design: Separate Names from Locations Packet forwarding – VL2 agent (at host) traps packets and encapsulates them Address resolution – ARP requests converted to unicast to directory system – Cached for performance Access control (security policy) via the directory system Directory System User space Kernel Server Machine ApplicationVL2 Agent LookUp (AA) IncapInfo (AA)

Design: Separate Names from Locations

Design : Valiant Load Balancing Each flow goes through a different random path Hot-spot free for tested TMs

Design : VL2 Directory System Built using servers from the data center Two-tiered directory system architecture – Tier 1 : read optimized cache servers (directory server) – Tier 2 : write optimized mapping servers (RSM)

Benefits + Drawbacks

Benefits VM migration – No need to worry L2 broadcast – Location+address dependence Revisiting fault tolerance – Placement requirements

Loop-free Forwarding and Fault-Tolerant Routing Switches build forwarding tables based on their position – edge, aggregation and core switches Use strict “up-down semantics” to ensure loop-free forwarding – Load-balancing: use any ECMP path via flow hashing to ensure packet ordering Fault-tolerant routing: – Mostly concerned with detecting failures – Fabric manager maintains logical fault matrix with per-link connectivity info; inform affected switches – Affected switches re-compute forwarding tables 24

Draw Backs Higher failures – Commodity switches fail more frequently No straight forward way to expand – Expand in large increments, values of k Look-up servers – Additional infrastructure servers – Higher upfront startup latency Need special gateway servers

Draw Backs Higher failures – Commodity switches fail more frequently No straight forward way to expand – Expand in large increments, values of k Look-up servers – Additional infrastructure servers – Higher upfront startup latency

Draw Backs Higher failures – Commodity switches fail more frequently No straight forward way to expand – Expand in large increments, values of k Look-up servers – Additional infrastructure servers – Higher upfront startup latency