Before We Start How to read a research paper?

Slides:



Advertisements
Similar presentations
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric
Advertisements

COS 461 Fall 1997 Routing COS 461 Fall 1997 Typical Structure.
CSCI 465 D ata Communications and Networks Lecture 20 Martin van Bommel CSCI 465 Data Communications & Networks 1.
PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. Presented by: Vinuthna Nalluri Shiva Srivastava.
Data Center Fabrics. Forwarding Today Layer 3 approach: – Assign IP addresses to hosts hierarchically based on their directly connected switch. – Use.
NETWORK LOAD BALANCING NLB.  Network Load Balancing (NLB) is a Clustering Technology.  Windows Based. (windows server).  To scale performance, Network.
Datacenter Network Topologies
1 Internet Networking Spring 2004 Tutorial 13 LSNAT - Load Sharing NAT (RFC 2391)
Subnetting.
PortLand Presented by Muhammad Sadeeq and Ling Su.
Data Center Network Topologies: FatTree
Delivery, Forwarding, and Routing
A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares, Alexander Loukissas, Amin Vahdat Presented by Gregory Peaker and Tyler Maclean.
TCP: Software for Reliable Communication. Spring 2002Computer Networks Applications Internet: a Collection of Disparate Networks Different goals: Speed,
Chapter 9 Classification And Forwarding. Outline.
COS 420 Day 16. Agenda Assignment 3 Corrected Poor results 1 C and 2 Ds Spring Break?? Assignment 4 Posted Chap Due April 6 Individual Project Presentations.
A Scalable, Commodity Data Center Network Architecture Mohammad AI-Fares, Alexander Loukissas, Amin Vahdat Presented by Ye Tao Feb 6 th 2013.
A Scalable, Commodity Data Center Network Architecture
Understanding Network Failures in Data Centers: Measurement, Analysis and Implications Phillipa Gill University of Toronto Navendu Jain & Nachiappan Nagappan.
A Scalable, Commodity Data Center Network Architecture.
TCP/IP Reference Model Host To Network Layer Transport Layer Application Layer Internet Layer.
Lecture Week 3 Introduction to Dynamic Routing Protocol Routing Protocols and Concepts.
Router Architectures An overview of router architectures.
(part 3).  Switches, also known as switching hubs, have become an increasingly important part of our networking today, because when working with hubs,
Router Architectures An overview of router architectures.
Network Topologies.
Introduction to IT and Communications Technology Justin Champion C208 – 3292 Ethernet Switching CE
Presentation Title Subtitle Author Copyright © 2002 OPNET Technologies, Inc. TM Introduction to IP and Routing.
1 Chapter 27 Internetwork Routing (Static and automatic routing; route propagation; BGP, RIP, OSPF; multicast routing)
University of the Western Cape Chapter 11: Routing Aleksandar Radovanovic.
1 Chapter 27 Internetwork Routing (Static and automatic routing; route propagation; BGP, RIP, OSPF; multicast routing)
 Network Segments  NICs  Repeaters  Hubs  Bridges  Switches  Routers and Brouters  Gateways 2.
A Scalable, Commodity Data Center Network Architecture Jingyang Zhu.
DARD: Distributed Adaptive Routing for Datacenter Networks Xin Wu, Xiaowei Yang.
CS1Q Computer Systems Lecture 17 Simon Gay. Lecture 17CS1Q Computer Systems - Simon Gay2 The Layered Model of Networks It is useful to think of networks.
Floodless in SEATTLE : A Scalable Ethernet ArchiTecTure for Large Enterprises. Changhoon Kim, Matthew Caesar and Jenifer Rexford. Princeton University.
LAN Switching and Wireless – Chapter 1 Vilina Hutter, Instructor
Department of Computer Science A Scalable, Commodity Data Center Network Architecture Mohammad Al-Fares Alexander Loukissas Amin Vahdat SIGCOMM’08 Reporter:
IP1 The Underlying Technologies. What is inside the Internet? Or What are the key underlying technologies that make it work so successfully? –Packet Switching.
Packet switching network Data is divided into packets. Transfer of information as payload in data packets Packets undergo random delays & possible loss.
S-Paxos: Eliminating the Leader Bottleneck
Netprog: Routing and the Network Layer1 Routing and the Network Layer (ref: Interconnections by Perlman)
Routing and Routing Protocols
Chapter 11 Extending LANs 1. Distance limitations of LANs 2. Connecting multiple LANs together 3. Repeaters 4. Bridges 5. Filtering frame 6. Bridged network.
Computer Science and Engineering Copyright by Hesham El-Rewini Advanced Computer Architecture CSE 8383 April 11, 2006 Session 23.
1 Version 3.1 Module 6 Routed & Routing Protocols.
Chapter 4 Version 1 Virtual LANs. Introduction By default, switches forward broadcasts, this means that all segments connected to a switch are in one.
Mapping IP Addresses to Hardware Addresses Chapter 5.
[Fat- tree]: A Scalable, Commodity Data Center Network Architecture
IP Routing table compaction and sampling schemes to enhance TCAM cache performance Author: Ruirui Guo a, Jose G. Delgado-Frias Publisher: Journal of Systems.
1 IP Routing table compaction and sampling schemes to enhance TCAM cache performance Author: Ruirui Guo, Jose G. Delgado-Frias Publisher: Journal of Systems.
MMPTCP: A Multipath Transport Protocol for Data Centres 1 Morteza Kheirkhah University of Edinburgh, UK Ian Wakeman and George Parisis University of Sussex,
1 K. Salah Module 5.1: Internet Protocol TCP/IP Suite IP Addressing ARP RARP DHCP.
Ethernet Packet Filtering - Part1 Øyvind Holmeide Jean-Frédéric Gauvin 05/06/2014 by.
Network Virtualization Ben Pfaff Nicira Networks, Inc.
Data Center Architectures
CIS 700-5: The Design and Implementation of Cloud Networks
Lecture 2: Cloud Computing
Data Center Network Topologies II
Data Center Network Architectures
The Underlying Technologies
ECE 544: Traffic engineering (supplement)
VIRTUAL SERVERS Presented By: Ravi Joshi IV Year (IT)
A Scalable, Commodity Data Center Network Architecture
CMPE 252A : Computer Networks
What’s “Inside” a Router?
NTHU CS5421 Cloud Computing
Data Center Architectures
CMPE 252A : Computer Networks
Routing and the Network Layer (ref: Interconnections by Perlman
Presentation transcript:

Before We Start How to read a research paper? How to write a paper review?

Reading A Research Paper Differs from reading a text book May not always provide all background details Iteratively activity Often requires obtaining additional background or skipping details

Active Reading Use a highlighter Write questions down Underline Go back and forth between sections

Organization Abstract Introduction Background/Related work Proposed Technique Evaluation Conclusion

Abstract Usually less than 300 words. Summarizes the paper including brief motivation, high level description of contribution and results/ conclusions Stands on its own. The abstract of a paper can be removed and the rest of the paper is still complete Don’t need to define all terms in an abstract If you do define them you need to re-define in proper part of paper

Introduction This provides the high level motivation for the work. If the introduction is not compelling the paper will not be read Should also state what problem is being solved, what other work has been done that is similar, and why this work is unique Often ends with a list of the contributions

Introduction App oversubscription Under client web service control VM Guest OS App Under client web service control oversubscription

Motivation The principle bottleneck in large-scale clusters is often inter-node communication bandwidth Two solutions: Specialized hardware and communication protocols e.g. Infiband, Myrinet(supercomputer environment) Cons: No commodity parts (expensive) The protocols are not compatible with TCP/IP Commodity Ethernet Switches Cons: Scale poorly non-linear cost increases with cluster Size. high-end core switch, oversubscription(tradeoff)

Oversubscription Ratio …………… Server 2 Server n B Upper Link Bandwidth(UB) B Server 1 Oversubscription Ratio= B*n/UB

Current Data Center Topology Edge hosts connect to 1G Top of Rack (ToR) switch ToR switches connect to 10G End of Row (EoR) switches Large clusters: EoR switches to 10G core switches Oversubscription of 2.5:1 to 8:1 typical in guidelines Key challenges: performance, cost, routing, energy, cabling

Data Center Cost

Design Goals Scalable interconnection bandwidth Arbitrary host communication at full bandwidth Economies of scale Commodity Switch Backward compatibility Compatible with hosts running Ethernet and IP

Related Work Sometimes combined with background Briefly describes other similar research papers and tells how the work presented in the current paper differs Not meant to be a complete summary of each paper Usually categorizes other work into like groups. Often found right before conclusions

Background Any technical background that is required to understand the techniques that are presented should appear here. Often has a motivating example that can be used to explain the ideas.

Proposed Technique This is where you provide detail of the new material in your paper. You can present algorithms/methods/processes, etc. If you have a running example from the intro or background you can use this to illustrate your ideas.

Fat-Tree Topology k/2 Aggregation Switch in each pod   K Pods k/2 Aggregation Switch in each pod k/2 Edge Switches in each pod k/2 servers in each Rack

Fat-tree Topology Equivalent

Routing (k/2)*(k/2) shortest path! IP needs extension here! Single-Path Routing VS Multi-Path Routing Static VS Dynamic

ECMP(Equal-Cost Multiple-Path Routing) Static Follow scheduling limited multiplicity of path to 8-16 Increase routing table multiplicatively, hence latency time Advantage: Don’t need reordering! Modern Switch support! Extract Source and Destination Address Hash Function(CRC16) Determine which region fall in 1 2 3 4   Hash-Threshold

Two-level Routing Table Routing Aggregation 192.168.1.2/24 192.168.1.10/24 192.168.1.45/24 192.168.1.89/24 1 192.168.2.3/24 192.168.1.0/24 192.168.2.0/24 1 192.168.2.8/24 192.168.2.10/24

Two-level Routing Table Addressing Using 10.0.0.0/8 private IP address Pod Switch: 10. pod. Switch.1. pod range is [0, k-1](left to right) switch range is [0, k-1] (left to right, bottom to top) Core Switch: 10. k. i . j (i,j) is the point in (k/2)*(k/2) grid Host: 10.pod. Switch.ID ID range is [2, k/2+1] (left to right)

Two-level Routing Table 10.0.0.1 10.0.0.2 10.0.0.3 10.4.1.1 10.4.1.2 10.4.2.1 10.4.2.2 10.2.0.2 10.2.0.3

Two-level Routing Table Two-level Routing Table Structure Two-level Routing Table implementation TCAM=Ternary Content-Addressable Memory Parallel searching Priority encoding

Two-level Routing Table---example Prefix Outgoing Port 10.0.0.0/24 10.0.1.1/24 1 0.0.0.0/0 Suffix Outgoing Port 0.0.0.2/8 3 0.0.0.3/8 2 Two-level Routing Table---example Prefix Outgoing Port 10.0.0.0/16 10.1.0.0/16 1 10.2.0.0/16 2 10.3.0.0/16 3 example Prefix Outgoing Port 10.2.0.0/24 10.2.1.0/24 1 0.0.0.0/0 Prefix Outgoing Port 10.2.0.2/32 10.2.1.3/32 1 0.0.0.0/0 Suffix Outgoing Port 0.0.0.2/8 2 0.0.0.3/8 3 Suffix Outgoing Port 0.0.0.2/8 2 0.0.0.3/8 3 Prefix Outgoing Port 10.0.1.2/32 10.0.1.3/32 1 0.0.0.0/0 Suffix Outgoing Port 0.0.0.2/8 2 0.0.0.3/8 3

Two-Level Routing Table Avoid Packet Reordering traffic diffusion occurs in the first half of a packet journey Centralized Protocol to Initialize the Routing Table

Flow Classification(Dynamic) Soft State(Compatible with Two-Level Routing Table) A flow=packet with the same source and destination IP address Avoid Reordering of Flow Balancing Assignment and Updating

Flow Classification—Flow Assignment Hash(Src,Des) Have seen this hash value? Lookup previously assign port x Send packet on port x Y Record new flow record f Assign f to least-loaded port x N

Flow Classification—Update  

Flow Scheduling distribution of transfer times and burst lengths of Internet traffic is long-tailed Large flow dominating Large flow should be specially handled

Flow Scheduling Notify the central controller Eliminates global congestion Prevent long lived flows from sharing the same links Assign long lived flows to different links Edge Switch Detecting Flow size above a threshold Notify the central controller Assign this flow to non-conflicting path

Fault-Tolerance Bidirectional Forwarding Detection session (BFD) Lower- to Upper-layer Switches Upper-layer to Core Switches For flow scheduling, it is much more easier to handle.

Failure b/w upper layer and core switches Outgoing inter-pod traffic: local routing table marks the affected link as unavailable and chooses another core switch Incoming inter-pod traffic: core switch broadcasts a tag to upper switches directly connected signifying its inability to carry traffic to that entire pod, then upper switches avoid that core switch when assigning flows destined to that pod

Failure b/w lower and upper layer switches Outgoing inter- and intra pod traffic from lower-layer: the local flow classifier sets the cost to infinity and does not assign it any new flows, chooses another upper layer switch Intra-pod traffic using upper layer switch as intermediary: Switch broadcasts a tag notifying all lower level switches, these would check when assigning new flows and avoid it Inter-pod traffic coming into upper layer switch: Tag to all its core switches signifying its ability to carry traffic, core switches mirror this tag to all upper layer switches, then upper switches avoid affected core switch when assigning new flaws

Evaluation A strong paper has very extensive evaluation, including testbed implementation based and simulation based evaluation. Parts of this section include: methdology, metrics used, results, comparison, etc

Experiment Description—Fat-tree, Click 4-port fat-tree, there are 16 hosts, four pods (each with four switches), and four core switches. We multiplex these 36 elements onto ten physical machines, interconnected by a 48-port ProCurve 2900 switch with 1 Gigabit Ethernet links. Each pod of switches is hosted on one machine; each pod’s hosts are hosted on one machine; and the two remaining machines run two core switches each. bandwidth-limited to 96Mbit/s to ensure that the configuration is not CPU limited. Each host generates a constant 96Mbit/s of outgoing traffic

Experiment Description—hierarchical tree,click four machines running four hosts each, and four machines each running four pod switches with one additional uplink The four pod switches are connected to a 4- port core switch running on a dedicated machine. 3.6:1 oversubscription on the uplinks from the pod switches to the core switch Each host generates a constant 96Mbit/s of outgoing traffic

Result

Power and Heat

End of the Paper Conclusions and Future Work: Summarize your results and any conclusions drawn Describe briefly the main areas you plan to pursue as future work

Conclusion Bandwidth is the scalability bottleneck in large scale clusters Existing solutions are expensive and limit cluster size Fat-tree topology with scalable routing and backward compatibility with TCP/IP and Ethernet

What Makes A Good Research Paper Good methodology/technique Novel: new and not resembling something formerly known or used Good writing Publish in good journals/conferences

Paper Reviews We will write a series of paper review in this class First review due on next Tuesday in class for this paper Hand in a hard copy

Paper Review Form Paper (author, title, complete reference) Short 100 word overview description Questions: What problem is the paper addressing? What is the contribution? Did you find any drawbacks? What is your assessment of the overall presentation style of the paper (consistency, clarity, ease of reading)? Any possible future work or improvements?