Chuanxiong Guo, Haitao Wu, Kun Tan,

Slides:



Advertisements
Similar presentations
Path Splicing with Network Slicing
Advertisements

Comparison Of Network On Chip Topologies Ahmet Salih BÜYÜKKAYHAN Fall.
Computer Network Topologies
Chuanxiong Guo, Haitao Wu, Kun Tan, Lei Shi, Yongguang Zhang, Songwu Lu Microsoft Research Asia, Tsinghua University, UCLA 1 DCell: A Scalable and Fault-Tolerant.
BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers Chuanxiong Guo1, Guohan Lu1, Dan Li1, Haitao Wu1, Xuan Zhang2,
J. K. Kim Portland. Trend of Data Center By J. Nicholas Hoover, InformationWeek June 17, :00 AMJ. Nicholas Hoover 200 million Euro Data centers.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. Presented by: Vinuthna Nalluri Shiva Srivastava.
Data Center Fabrics. Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Emery Berger University of Massachusetts Amherst Operating Systems CMPSCI 377 Lecture.
Datacenter Network Topologies
1 Lecture 23: Interconnection Networks Topics: communication latency, centralized and decentralized switches (Appendix E)
1 13-Jun-15 S Ward Abingdon and Witney College LAN design CCNA Exploration Semester 3 Chapter 1.
presented by Hasan SÖZER1 Scalable P2P Search Daniel A. Menascé George Mason University.
BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers Chuanxiong Guo1, Guohan Lu1, Dan Li1, Haitao Wu1, Xuan Zhang2,
Portland: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric Offense Kai Chen Shih-Chi Chen.
Department of Computer Science, Jinan University, Guangzhou, P.R. China Lijun Lyu, Junjie Xie, Yuhui Deng, Yongtao Zhou ICA3PP 2014: The 14th International.
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
Ji-Yong Shin * Bernard Wong +, and Emin Gün Sirer * * Cornell University + University of Waterloo 2 nd ACM Symposium on Cloud ComputingOct 27, 2011 Small-World.
A Scalable, Commodity Data Center Network Architecture Mohammad AI-Fares, Alexander Loukissas, Amin Vahdat Presented by Ye Tao Feb 6 th 2013.
A Scalable, Commodity Data Center Network Architecture
A Scalable, Commodity Data Center Network Architecture.
Switching, routing, and flow control in interconnection networks.
B 謝宗廷 B96b02016 張煥基 1. Outline Introduction Bcube structure Bcube source routing OTHER DESIGN ISSUES GRACEFUL DEGRADATION Implementation Architecture.
CS An Overlay Routing Scheme For Moving Large Files Su Zhang Kai Xu.
Presenter: Po-Chun Wu. Outline Introduction BCube Structure BCube Source Routing (BSR) Other Design Issues Graceful degradation Implementation.
Routing & Architecture
1 Department of Computer Science, Jinan University 2 School of Computer Science & Technology, Huazhong University of Science & Technology Junjie Xie 1,
A Scalable, Commodity Data Center Network Architecture Jingyang Zhu.
DARD: Distributed Adaptive Routing for Datacenter Networks Xin Wu, Xiaowei Yang.
LAN Switching and Wireless – Chapter 1
1 LAN design- Chapter 1 CCNA Exploration Semester 3 Modified by Profs. Ward and Cappellino.
Floodless in SEATTLE : A Scalable Ethernet ArchiTecTure for Large Enterprises. Changhoon Kim, Matthew Caesar and Jenifer Rexford. Princeton University.
A Routing Underlay for Overlay Networks Akihiro Nakao Larry Peterson Andy Bavier SIGCOMM’03 Reviewer: Jing lu.
LAN Switching and Wireless – Chapter 1 Vilina Hutter, Instructor
VL2: A Scalable and Flexible Data Center Network Albert Greenberg, James R. Hamilton, Navendu Jain, Srikanth Kandula, Changhoon Kim, Parantap Lahiri, David.
Department of Computer Science A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares Alexander Loukissas Amin Vahdat SIGCOMM’08 Reporter:
InterConnection Network Topologies to Minimize graph diameter: Low Diameter Regular graphs and Physical Wire Length Constrained networks Nilesh Choudhury.
Dual Centric Data Center Network Architectures DAWEI LI, JIE WU (TEMPLE UNIVERSITY) ZHIYONG LIU, AND FA ZHANG (CHINESE ACADEMY OF SCIENCES) ICPP 2015.
Chuanxiong Guo, Haitao Wu, Kun Tan, Lei Shi, Yongguang Zhang, Songwu Lu SIGCOMM 2008 Presented by Ye Tian for Course CS05112.
Subways: A Case for Redundant, Inexpensive Data Center Edge Links Vincent Liu, Danyang Zhuo, Simon Peter, Arvind Krishnamurthy, Thomas Anderson University.
SYMBIOTIC ROUTING IN FUTURE DATA CENTER 工科三 陳泰穎. Outline 1. CamCube 1. Traditional data center 2. The problems 3. CamCube philosophy 4. Feature 5. What’s.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
SecondNet: A Data Center Network Virtualization Architecture with Bandwidth Guarantees Chuanxiong Guo 1, Guohan Lu 1, Helen J. Wang 2, Shuang Yang 3, Chao.
Jiaxin Cao, Rui Xia, Pengkun Yang, Chuanxiong Guo,
Data Centers and Cloud Computing 1. 2 Data Centers 3.
BUFFALO: Bloom Filter Forwarding Architecture for Large Organizations Minlan Yu Princeton University Joint work with Alex Fabrikant,
VL2: A Scalable and Flexible Data Center Network
Yiting Xia, T. S. Eugene Ng Rice University
Instructor Materials Chapter 1: LAN Design
CIS 700-5: The Design and Implementation of Cloud Networks
Architecture and Algorithms for an IEEE 802
Data Center Network Architectures
Networks Network:end-node and router C 2 B 1 3 D 5 A 4 6 E 7 Router F
Chuanxiong Guo, et al, Microsoft Research Asia, SIGCOMM 2008
Computer Network Topologies
FAR: A Fault-avoidance Routing Method for Data Center Networks with Regular Topology Please send.
NTHU CS5421 Cloud Computing
A Scalable, Commodity Data Center Network Architecture
IS3120 Network Communications Infrastructure
BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers Chuanxiong Guo1, Guohan Lu1, Dan Li1, Haitao Wu1, Xuan Zhang2,
Early Measurements of a Cluster-based Architecture for P2P Systems
Chuanxiong Guo, Haitao Wu, Kun Tan,
Generic and Automatic Address Configuration for Data Center Networks
Generic and Automatic Address Configuration for Data Center Networks
NTHU CS5421 Cloud Computing
VL2: A Scalable and Flexible Data Center Network
Jellyfish: Networking Data Centers Randomly
RDMA over Commodity Ethernet at Scale
Data Center Architectures
In-network computation
Presentation transcript:

DCell: A Scalable and Fault Tolerant Network Structure for Data Centers Chuanxiong Guo, Haitao Wu, Kun Tan, Lei Shi, Yongguang Zhang, Songwu Lu Wireless and Networking Group Microsoft Research Asia August 19, 2008, ACM SIGCOMM

Outline DCN motivation DCell Routing in DCell Simulation Results Implementation and Experiments Related work Conclusion

Data Center Networking (DCN) Ever increasing scale Google has 450,000 servers in 2006 Microsoft doubles its number of servers in 14 months The expansion rate exceeds Moore’s Law Network capacity: Bandwidth hungry data-centric applications Data shuffling in MapReduce/Dryad Data replication/re-replication in distributed file systems Index building in Search Fault-tolerance: When data centers scale, failures become the norm Cost: Using high-end switches/routers to scale up is costly

Interconnection Structure for Data Centers Existing tree structure does not scale Expensive high-end switches to scale up Single point of failure and bandwidth bottleneck Experiences from real systems ? Our answer: DCell

DCell Ideas #1: Use mini-switches to scale out #2: Leverage servers be part of the routing infrastructure Servers have multiple ports and need to forward packets #3: Use recursion to scale and build complete graph to increase capacity

DCell: the Construction n=2, k=2 DCell_1 n=2, k=1 Dcell_0 Server Mini-switch n servers in a DCell_0 n=2, k=0 End recursion by building DCell0 Build sub-DCells Connect sub-DCells to form complete graph

DCell: The Properties Scalability: The number of servers scales doubly exponentially Where number of servers in a DCell0 is 8 (n=8) and the number of server ports is 4 (i.e., k=3) -> N=27,630,792 Fault-tolerance: The bisection width is larger than No severe bottleneck links: Under all-to-all traffic pattern, the number of flows in a level-i link is less than For tree, under all-to-all traffic pattern, the max number of flows in a link is in proportion to

Routing without Failure: DCellRouting src n1 n2 dst Time complexity: 2k+1 steps to get the whole path k+1 to get the next hop

DCellRouting (cont.) Network diameter: The maximum path length using DCellRouting in a DCellk is at most But: DCellRouting is NOT a shortest-path routing is NOT a tight diameter bound for DCell The mean and max path lengths of shortest-path and DCellRouting n k N Shortest-path DCellRouting Mean Max 4 2 420 4.87 7 5.16 5 930 5.22 5.50 6 1806 5.48 5.73 3 176,820 9.96 15 11.29 865,830 10.74 11.98 3,263,442 11.31 12.46 Yet: DCellRouting is close to shortest-path routing DCellRouting is much simpler: O(k) steps to decide the next hop

DFR: DCell Fault-tolerant Routing Design goal: Support millions of servers Advantages to take: DCellRouting and DCell topology Ideas #1: Local-reroute and Proxy to bypass failed links Take advantage of the complete graph topology #2: Local Link-state To avoid loops with only local-reroute #3: Jump-up for rack failure To bypass a whole failed rack

DFR: DCell Fault-tolerant Routing src dst m1 m2 n2 n1 r1 DCellb i1 i2 L p1 q2 i3 DCellb p2 q1 Proxy L Proxy L+1 s2 s1 Servers in a same share local link-state

DFR Simulations: Server failure Two DCells: n=4, k=3 -> N=176,820 n=5, k=3 -> N=865,830

DFR Simulations: Rack failure Two DCells: n=4, k=3 -> N=176,820 n=5, k=3 -> N=865,830

DFR Simulations: Link failure Two DCells: n=4, k=3 -> N=176,820 n=5, k=3 -> N=865,830

DCN (routing, forwarding, address mapping, ) Implementation DCell Protocol Suite Design Apps only see TCP/IP Routing is in DCN (IP addr can be flat) Software implementation A 2.5 layer approach Use CPU for packet forwarding Next: Offload packet forwarding to hardware APP TCP/IP DCN (routing, forwarding, address mapping, ) Ethernet Intel® PRO/1000 PT Quad Port Server Adapter

Testbed DCell1: 20 servers, 5 DCell0s DCell0: 4 servers Ethernet wires 8-port mini-switches, 50$ each

Fault Tolerance DCell fault-tolerant routing can handle various failures Link failure Server/switch failure Rack failure Link failure Server shutdown

Network Capacity All to all traffic: each server sends 5GB file to every other servers

Related Work Hypercube: node degree is large Butterfly and FatTree: scalability is not as fast as DCell De Bruijn: cannot incrementally expand

Related Work

Summary DCell: Benefits: Use commodity mini-switches to scale out Let (NIC of) servers be part of the routing infrastructure Use recursion to reduce the node degree and complete graph to increase network capacity Benefits: Scales doubly exponentially High aggregate bandwidth capacity Fault tolerance Cost saving Ongoing work: move packet forwarding into FPGA Price to pay: higher wiring cost In summary, we have presented dcell, the fault-tolerant routing protocol on top of it, simulations and testbed experiments to demonstrates the performance of dcell. One price to pay in DCell, as well as in other low dimensional structures, is much higher wiring cost.

I would like to give you evidence that wiring has been well addressed in other communities. The bird nest, our splendid national stadium for the ongoing Olympic games is weaved together by many, many long wires!

Q & A That’s the end my presentation. Thank you. Any questions?