Routing & Architecture

Slides:



Advertisements
Similar presentations
Chuanxiong Guo, Haitao Wu, Kun Tan, Lei Shi, Yongguang Zhang, Songwu Lu Microsoft Research Asia, Tsinghua University, UCLA 1 DCell: A Scalable and Fault-Tolerant.
Advertisements

Communication Networks Recitation 3 Bridges & Spanning trees.
University of Calgary – CPSC 441.  We need to break down big networks to sub-LANs  Limited amount of supportable traffic: on single LAN, all stations.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric
BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers Chuanxiong Guo1, Guohan Lu1, Dan Li1, Haitao Wu1, Xuan Zhang2,
Multicast in Wireless Mesh Network Xuan (William) Zhang Xun Shi.
Fundamentals of Computer Networks ECE 478/578 Lecture #13: Packet Switching (2) Instructor: Loukas Lazos Dept of Electrical and Computer Engineering University.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. Presented by: Vinuthna Nalluri Shiva Srivastava.
Radhika Niranjan Mysore, Andreas Pamboris, Nathan Farrington, Nelson Huang, Pardis Miri, Sivasankar Radhakrishnan, Vikram Subramanya, and Amin Vahdat Department.
Data Center Fabrics. Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use.
1 Exploring Efficient and Scalable Multicast Routing in Future Data Center Networks Dan Li, Jiangwei Yu, Junbiao Yu, Jianping Wu Tsinghua University Presented.
Datacenter Network Topologies
5/31/05CS118/Spring051 twisted pair hub 10BaseT, 100BaseT, hub r T= Twisted pair (copper wire) r Nodes connected to a hub, 100m max distance r Hub: physical.
A Comparison of Layering and Stream Replication Video Multicast Schemes Taehyun Kim and Mostafa H. Ammar.
CS 268: Lecture 5 (Project Suggestions) Ion Stoica February 6, 2002.
1 Computer Networks Internetworking Devices. 2 Repeaters Hubs Bridges –Learning algorithms –Problem of closed loops Switches Routers.
Communication operations Efficient Parallel Algorithms COMP308.
CS 268: Project Suggestions Ion Stoica February 6, 2003.
BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers Chuanxiong Guo1, Guohan Lu1, Dan Li1, Haitao Wu1, Xuan Zhang2,
Portland: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric Offense Kai Chen Shih-Chi Chen.
Chuanxiong Guo, Haitao Wu, Kun Tan,
Tesseract A 4D Network Control Plane
PortLand Presented by Muhammad Sadeeq and Ling Su.
Data Center Network Topologies: FatTree
Department of Computer Science, Jinan University, Guangzhou, P.R. China Lijun Lyu, Junjie Xie, Yuhui Deng, Yongtao Zhou ICA3PP 2014: The 14th International.
Introduction to Computer Networks 09/23 Presenter: Fatemah Panahi.
Ji-Yong Shin * Bernard Wong +, and Emin Gün Sirer * * Cornell University + University of Waterloo 2 nd ACM Symposium on Cloud ComputingOct 27, 2011 Small-World.
A Scalable, Commodity Data Center Network Architecture Mohammad AI-Fares, Alexander Loukissas, Amin Vahdat Presented by Ye Tao Feb 6 th 2013.
A Scalable, Commodity Data Center Network Architecture
Lawrence G. Roberts CEO Anagran September 2005 Advances Toward Economic and Efficient Terabit LANs and WANs.
B 謝宗廷 B96b02016 張煥基 1. Outline Introduction Bcube structure Bcube source routing OTHER DESIGN ISSUES GRACEFUL DEGRADATION Implementation Architecture.
Transparent Bridging. Chapter Goals Understand transparent bridge processes of learning, filtering, forwarding, and flooding. Explain the purpose of the.
TRansparent Interconnection of Lots of Links (TRILL) March 11 th 2010 David Bond University of New Hampshire: InterOperability.
Presenter: Po-Chun Wu. Outline Introduction BCube Structure BCube Source Routing (BSR) Other Design Issues Graceful degradation Implementation.
VL2 – A Scalable & Flexible Data Center Network Authors: Greenberg et al Presenter: Syed M Irteza – LUMS CS678: 2 April 2013.
1 Department of Computer Science, Jinan University 2 School of Computer Science & Technology, Huazhong University of Science & Technology Junjie Xie 1,
 Network Segments  NICs  Repeaters  Hubs  Bridges  Switches  Routers and Brouters  Gateways 2.
1 Next Few Classes Networking basics Protection & Security.
LAN Switching and Wireless – Chapter 1
Network and Communications Ju Wang Chapter 5 Routing Algorithm Adopted from Choi’s notes Virginia Commonwealth University.
Floodless in SEATTLE : A Scalable Ethernet ArchiTecTure for Large Enterprises. Changhoon Kim, Matthew Caesar and Jenifer Rexford. Princeton University.
Link Layer Review CS244A Winter 2008 March 7, 2008 Ben Nham.
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
Department of Computer Science A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares Alexander Loukissas Amin Vahdat SIGCOMM’08 Reporter:
Networking Fundamentals. Basics Network – collection of nodes and links that cooperate for communication Nodes – computer systems –Internal (routers,
Chuanxiong Guo, Haitao Wu, Kun Tan, Lei Shi, Yongguang Zhang, Songwu Lu SIGCOMM 2008 Presented by Ye Tian for Course CS05112.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
Winter 2008CS244a Handout 111 CS244a: An Introduction to Computer Networks Handout 11: Interconnecting LANs Nick McKeown Professor of Electrical Engineering.
Approaches to Improve Data Center Performance through Networking - Gurubaran.
VL2: A Scalable and Flexible Data Center Network
Data Center Architectures
Performance and Energy Efficiency Metrics for Communication Systems of Cloud Computing Data Centers Hrushikesh Mahapatro IT
Yiting Xia, T. S. Eugene Ng Rice University
Data Center Networking
Data Center Network Topologies II
Data Center Network Architectures
Chuanxiong Guo, et al, Microsoft Research Asia, SIGCOMM 2008
Advanced Computer Networks
: An Introduction to Computer Networks
NTHU CS5421 Cloud Computing
A Scalable, Commodity Data Center Network Architecture
BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers Chuanxiong Guo1, Guohan Lu1, Dan Li1, Haitao Wu1, Xuan Zhang2,
Chuanxiong Guo, Haitao Wu, Kun Tan,
NTHU CS5421 Cloud Computing
ECE 544 Project3 Team member: BIAO LI, BO QU, XIAO ZHANG 1 1.
The Network Layer Network Layer Design Issues:
VL2: A Scalable and Flexible Data Center Network
Data Center Architectures
Bridges Neil Tang 10/10/2008 CS440 Computer Networks.
Data Center Traffic Engineering
Presentation transcript:

Routing & Architecture Comparison about: DCell , PortLand , VL2 , BCube , MDCube Point of views: Goals , artecheture , routing algorithm , fault tolerance , broadcast and multicast , evaluation and compare,

Goals Comparison about: DCell , PortLand , VL2 , BCube , MDCube

Goals DCell PortLand VL2 Scalable + incremental expansion VM migration Support high huge data centers with uniform high capacity between servers Fault-tolerance Adminstrator not need to configure any switch Performances isolation between services High network capacity Any end host should efficiently communicate with others. Ethernet layer-2 semantic No forwarding loop Failure detection

Goals BCube MDCube An architecture model for modular data center (MDC). Solve the scalability problem on BCube with small cost by increasing the diameter and some change of routing algorithm. High performance. For the same number of nodes, cost is significantly lower than BCube. Robust, graceful performance degradation for both switch/server faults

Artecheture Comparison about: DCell , PortLand , VL2 , BCube , MDCube

artecheture DCell

artecheture PortLand

artecheture VL2

artecheture DCell PortLand VL2 artecheture cost high wiring cost (compare the expensive core switch) --- diameter 2^(k+1) – 1 6 artecheture

artecheture BCube Recursively define architecture

artecheture MDCube Base on the architecture of BCube and some high speed fiber link. Each container is a BCube, and connect them with fiber. Usually one or two dimensions.

artecheture MDCube

artecheture BCube MDCube artecheture cost Only need cheap switches, but need a lot of switches when the number of nodes is very huge. With the help of fiber, the number of switches is much lower than BCube. If the level of BCube is not high( 2 or 3 ), then the cost will be low even compare to other architecture (FatTree, DCell, etc.) diameter 2* log p n p = number of ports in a switch n = number of nodes n= p^(k+1) k = level of Bcube 4k+3+D(2k+3) k is the level of Bcube D is the level of MDCube. artecheture

artecheture Diameter Compare Architecture Diameter Balance Tree 2k Fat-Tree 6 DCell 2^(k+1) – 1 BCube K + 1

Routing Algorithm Comparison about: DCell , PortLand , VL2 , BCube , MDCube

routing algorithm DCell PortLand VL2 Divide-and-conquer Two nodes src and dst that are in the same DCell(k) but in two different DCell(k-1)s. When computing the path from src to dst in a DCell k, we first calculate the intermediate link (n1; n2) that interconnects the two DCell(k-1)s. Just as TCP Agent add the header routing algorithm

DCell routing algorithm Divide-and-conquer PortLand VL2 Divide-and-conquer Two nodes src and dst that are in the same DCell(k) but in two different DCell(k-1)s. When computing the path from src to dst in a DCell k, we first calculate the intermediate link (n1; n2) that interconnects the two DCell(k-1)s. routing algorithm

DCell PortLand VL2 Just as TCP routing algorithm

DCell PortLand VL2 Agent add the header routing algorithm

routing algorithm BCube MDCube Each packet header contains the routing path of the packet, the relay node don’t have to decide the next node. One-to-one: One-to-many: All-to-all: One-to-all:

BCube routing algorithm MDCube Each packet header contains the routing path of the packet, the relay node don’t have to decide the next node. One-to-one: For node A = (a0,…,ak), B = (b0,…,bk), we can route A to B by changing one bit of the ID each time. Since the ID of the node is at most k+1, the length of the routing path is at most 2( k+1). There are k+1 parallel path. One-to-many: If node A want to send the data to x nodes, and x <= number of ports p, then we can construct a complete graph on these x+1 (including A) nodes by another x nodes as relays. One-to-all: Can build (k+1) edge disjoint trees, and use them to broadcast All-to-all: Can reach a good Aggregate Bottleneck Throughput (ABT) which is Θ(n) routing algorithm

MDCube routing algorithm BCube One-to-one: In the container, the routing algorithm is the same as BCube. When the path cross the container, correct the container ID one-by-one. The hops needed to intermediate the containers are 2D(k+1) + (D+1), D+1 is for the high speed link. This number plus the hop need in the two ends, which is 2*2(k+1), is the maximum hop needed in one routing path. There are k+1 parallel path. All-to-all: Similar to BCube, MDCube has ABT =Θ(n) routing algorithm

Fault Tolerance Comparison about: DCell , PortLand , VL2 , BCube , MDCube

fault tolerance DCell Local-reroute and Proxy, Local Link-state, Jump-up for Rack Failure PortLand Multi-rooted tree VL2 Completely bipartite BCube Yes, and has a good result on both switch and server fault. MDCube Good because of BCube.

fault tolerance - BCube

fault tolerance - BCube

Broadcast & Multicast Comparison about: DCell , PortLand , VL2 , BCube , MDCube

broadcast and multicast DCell : a sender delivers the broadcast packet to all its k + 1 neighbors when broadcasting a packet in a DCell(k). PortLand : nothing special VL2 : not mentioned BCube: Both can be reached. Do broadcast by constructing edge disjoint trees. Do multicast by using some relay nodes. MDCube: Similar to BCube.

Evalution & Compare Comparison about: DCell , PortLand , VL2 , BCube , MDCube

evalution and compare DCell : 1. The link failure incurs only 1-second throughput degradation, while the node failure incurs a 5- second throughput outage that corresponds to our link-state timeout value. 2. The aggregate throughput of DCell was about 2 times faster than Tree.

evalution and compare PortLand : Convergence time : 70ms (TCP : 200ms) Scalability : 100ARPs/sec/host 30000 nodes  control traffic 400Mbps; Core required : 70 VM migration : throughput drops to zero in 200-600ms, fully recovered after 32s

evalution and compare VL2 : Uniform high capacity VLB fairness Performance Isolation Directory-system performance

evalution and compare BCube:

evalution and compare BCube:

evalution and compare MDCube:

evalution and compare MDCube:

evalution and compare MDCube: