Presentation is loading. Please wait.

Presentation is loading. Please wait.

Lecture 9, Computer Networks (198:552)

Similar presentations


Presentation on theme: "Lecture 9, Computer Networks (198:552)"— Presentation transcript:

1 Lecture 9, Computer Networks (198:552)
Data Centers Part III Lecture 9, Computer Networks (198:552) Fall 2019 Slide material heavily adapted courtesy of Albert Greenberg, Changhoon Kim, Mohammad Alizadeh.

2 Data center costs Amortized Cost* Component Sub-Components ~45%
Servers CPU, memory, disk ~25% Power infrastructure UPS, cooling, power distribution ~15% Power draw Electrical utility costs Network Switches, links, transit Source: The Cost of a Cloud: Research Problems in Data Center Networks. Sigcomm CCR Greenberg, Hamilton, Maltz, Patel. *3 yr amortization for servers, 15 yr for infrastructure; 5% cost of money

3 Goal: Agility -- any service, any server
Turn the servers into a single large fungible pool Dynamically expand and contract service footprint as needed Place workloads where server resources are available Easier to maintain availability If one rack goes down, machines from another still available Want to view DCN as a pool of compute connected by one big high-speed fabric

4 How the network can support agility
Need flexibility & low configuration burden of L2 network Without the scaling concerns and single path routing restriction Need massive bisection bandwidth Limit oversubscription at higher layers of the tree topology

5 Building a high-speed switching fabric

6 A single (n X m)-port switching fabric
Different designs of switching fabric possible Assume n ingress ports and m egress ports, half duplex links

7 A single (n X m)-port switching fabric
Electrical/mechanical/electronic crossover We are OK with any design such that: Any port can connect to any other directly if all other ports free Nonblocking: if input port x and output port y are both free, they should be able to connect Regardless of other ports being connected. If not satisfied, switch is blocking.

8 High port density + nonblocking == hard!
Low-cost nonblocking crossbars are feasible for small # ports However, it is costly to be nonblocking with a large number of ports If each crossover is as fast as each input port, Number of crossover points == n * m Cost grows quadratically on the number of input ports Else, crossover must transition faster than the port … so that you can keep the number of crossovers small Q: How is this relevant to the data center network fabric?

9 Nonblocking switches with many ports
Key principle: Every fast nonblocking switch with a large number of ports is built out of many fast nonblocking switches with a small number of ports. How to build large nonblocking switches? The subject of interconnection networks from the telephony era

10 3-stage Clos network (r*n X r*n ports)
n X m r X r m X n n X m r X r m X n n X m r X r m X n r switches n X m m switches r X r r switches m X n

11 How Clos networks become nonblocking
if m > 2n – 2, then the Clos network is strict-sense nonblocking. That is, any new demand between any pair of free (input, output) ports can be satisfied without re-routing any of the existing demands. How?

12 Need at most (n-1)+(n-1) middle stage
n X m r X r m X n n X m r X r m X n At most n-1 existing demands n X m r X r m X n r switches n X m m switches r X r r switches m X n

13 Surprising result about Clos networks
if m >= n, then the Clos network is rearrangeably nonblocking That is, any new demand between any pair of free (input, output) ports can be satisfied by suitably re-routing existing demands. It is easy to see that m >= n is necessary The surprising part is that m >= n is sufficient

14 Rearrangeably nonblocking Clos built with identical switches
n X n n X n n X n n X n n X n n X n n X n n X n n X n n switches n X n n switches n X n n switches n X n

15 Modern data center network topologies are just folded Clos topologies.

16 How does one design a Clos DCN?
Switches are usually n X n with full-duplex links Fold the 3-stage Clos into 2-stages Share physical resources between ingress and egress stages Share ports and links across the two “sides” of the middle stage

17 … … … ToR/aggregation layer Spine layer ToR/aggregation layer
n/2 X n/2 n X n n/2 X n/2 From/to hosts n/2 X n/2 n X n n/2 X n/2 n/2 X n/2 n X n n/2 X n/2 n X n full duplex n/2 switches n X n full duplex n X n full duplex

18 Consequences of using folded Clos
2-stage high throughput data center topology All can use the same switches! (port density and link rates)

19 What about routing? We said that the Clos topology above is rearrangeably nonblocking. So, how to rearrange existing demands when a new packet arrives, so that it can get across as quickly as possible? How to do it without “interference” to (ie: rerouting) other pkts? VL2: We don’t need to rearrange anything.

20 Valiant Load Balancing (VLB)
Designed to move data quickly for shuffling in parallel computing Setting: Connectivity is sparse (“hypercube” topology): log n links per node in a network with n nodes Key idea: pick a random node to redirect a message to from the source, then follow the shortest path to the destination from there Guarantee: With high probability, the message reaches its destination very quickly (log n steps) Practically, this means there is very less queueing in the network

21 VLB in data center networks
VLB is more general than data center networks or Clos It is a form of oblivious routing No need to measure DCN traffic patterns before deciding on routing Extremely simple to implement: no global state VLB is very handy in folded Clos topologies due to the numerous options to pick the first-hop from the ToR switch. Balance load across many paths Very beneficial in practice Performance isolation: other flows don’t matter High capacity (“bisection bandwidth”) between any two ToR ports

22 Is ToR-port to ToR-port high capacity enough?
VLB + Clos provides high capacity if no ToR port demands more than its bandwidth. But is that enough?

23 Example: Web Search TLA MLA Worker Nodes ……… Deadline = 250ms Deadline = 50ms Deadline = 10ms A bunch of results arrive at MLAs and TLAs in a short period. Demand may exceed ToR port bandwidth! After the animation, say “this follows what we call the partition/aggregate application structure. I’ve described this in the context of search, but it’s actually quite general - “the foundation of many large-scale web applications.”

24 Hose traffic model The guarantees of VLB + Clos only hold under the hose model: Demands for any one ToR port (send or receive) must not exceed its bandwidth. Very hard to enforce especially on the receiver side without sender-side rate limits. VL2 uses TCP convergence as a way of ensuring that aggregate ToR port demand is within its bandwidth. Q: do you see any problems?

25 Discussion of VL2

26 What you said

27 Cost "In order to implement VL2, extra servers are needed, which brings more cost on devices. It seems like a tradeoff for a network with less over-subscription. It is not suitable to utilize VL2 in a large network because as the required servers increase, it becomes much much harder to implement than traditional data center network system and since the link and switches work all the time, it will cost more power. " Junlin Lu

28 Performance The requirement for the directory system is demanding. Is this system deployed in one single machine, or in a distributed system. If it comes to the latter, what is the network topology inside that system. In other words, what kind of distributed system can form a high-performing directory system to manage a high-performing distributed DCN." Huixiong Qin

29 Performance The paper proposes the addition of a shim layer which would fetch directory listing at the start of the flow. This process would be performed for each new flow and thus might affect performance. If the cluster is supporting highly interactive workloads such that they produce a large number of short-lived connections, the performance impact might be significant. Sukumar Gaonkar

30 Applicability and Context
The authors state at the beginning of the paper that they did not consider solutions that would require significant changes to the APIs of switches or similar network components. Had the authors considered this, are there any parts of their design that they would have changed significantly? Would considering such solutions only improve existing solutions slightly, or does it open a whole new array of ideas and solutions that were previously completely eliminated? Jakob Degen

31 Weaknesses One technical weakness I can think of is the lack of visibility that arises from using VLB. While the average case seems to be sufficient to ensure that hot spots do not arise in the network, there may be certain cases in which you want to see how flows are being routed or even control their route in particular as we discussed during the lecture on SDNs. In this case you have absolutely no control over how things are routed because it is chosen randomly. Amber Rawson

32 Weaknesses I am not sure whether the scalability of VL2 is good enough. Nowadays, the data centers are usually with hundreds of thousands of servers. Since the VL2 already shows its tail latency in a relative small datacenter environment, will a larger datacenter make the latency increase significantly? The maintaining of data consistency will also become harder if the network administrators want the DS to have small response delay. Qiongwen Xu

33 Building on VL2 Though the traffic pattern showed in the paper is totally volatile, I think these data can be applied to machine learning method, to see whether there is some pattern that can be configured with AI technology. Wenhao Lu

34

35 VL2: Virtual Layer 2 Backup slides

36 The Illusion of a Huge L2 Switch 3. Performance isolation
VL2 goals The Illusion of a Huge L2 Switch 1. L2 semantics 2. Uniform high capacity 3. Performance isolation A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A

37 VL2 design principles Randomizing to cope with volatility
Tremendous variability in traffic matrices Separating names from locations Any server, any service Embracing end systems Leverage the programmability & resources of servers Avoid changes to switches Building on proven networking technology Build with parts shipping today Leverage low cost, powerful merchant silicon ASICs, though do not rely on any one vendor

38 Single-chip “merchant silicon” switches
Switch ASIC 6 pack Wedge Image courtesy of Facebook

39 Specific objectives and solutions
Approach Solution 1. Layer-2 semantics Employ flat addressing Name-location separation & resolution service 2. Uniform high capacity between servers Guarantee bandwidth for hose-model traffic Flow-based random traffic indirection (Valiant LB) 3. Performance Isolation Enforce hose model using existing mechanisms only TCP

40 Addressing and routing: Name-location separation
Cope with host churns with very little overhead VL2 Switches run link-state routing and maintain only switch-level topology Directory Service x  ToR2 y  ToR3 z  ToR3 x  ToR2 y  ToR3 z  ToR4 ToR1 . . . ToR2 . . . ToR3 . . . ToR4 ToR3 y payload Lookup & Response x y y, z z ToR4 ToR3 z z payload payload Servers use flat names

41 Cope with host churns with very little overhead
Addressing and routing: Name-location separation Cope with host churns with very little overhead VL2 Switches run link-state routing and maintain only switch-level topology Directory Service Allows to use low-cost switches Protects network and hosts from host-state churn Obviates host and switch reconfiguration x  ToR2 y  ToR3 z  ToR4 x  ToR2 y  ToR3 z  ToR3 ToR1 . . . ToR2 . . . ToR3 . . . ToR4 ToR3 y payload Lookup & Response x y y, z z ToR4 ToR3 z z payload payload Servers use flat names

42 Example topology: Clos network
Offer huge aggr capacity and multi paths at modest cost VL2 Int . . . Aggr . . . K aggr switches with D ports . . . TOR . . . 20 Servers 20*(DK/4) Servers

43 Offer huge aggr capacity and multi paths at modest cost
Example topology: Clos network Offer huge aggr capacity and multi paths at modest cost VL2 Int . . . D (# of 10G ports) Max DC size (# of Servers) 48 11,520 96 46,080 144 103,680 Aggr . . . K aggr switches with D ports . . . TOR . . . 20 Servers 20*(DK/4) Servers

44 Traffic forwarding: Random indirection
Cope with arbitrary TMs with very little overhead Links used for up paths Links used for down paths IANY IANY IANY T1 T2 T3 T4 T5 T6 IANY T5 T3 z y payload payload x y z

45 Cope with arbitrary TMs with very little overhead
Traffic forwarding: Random indirection Cope with arbitrary TMs with very little overhead Links used for up paths Links used for down paths IANY IANY IANY [ ECMP + IP Anycast ] Harness huge bisection bandwidth Obviate esoteric traffic engineering or optimization Ensure robustness to failures Work with switch mechanisms available today T1 T2 T3 T4 T5 T6 IANY T3 T5 z y payload payload x y z

46 Some other DC network designs…
BCube [SIGCOMM’10] Jellyfish (random) [NSDI’12]


Download ppt "Lecture 9, Computer Networks (198:552)"

Similar presentations


Ads by Google