Download presentation
Presentation is loading. Please wait.
1
1 Architectural Results in the Optical Router Project Da Chuang, Isaac Keslassy, Nick McKeown High Performance Networking Group http://klamath.stanford.edu
2
2 Internet traffic x2/yr Router capacity x2.2/18 months 5x
3
3 POP with smaller routersPOP with large routers Interfaces: Price >$200k, Power > 400W About 50-60% of interfaces are used for interconnection within the POP. Industry trend is towards large, single router per POP. Fast (large) routers Big POPs need big routers
4
4 100Tb/s optical router Objective To determine the best way to incorporate optics into routers. Push technology hard to expose new issues. Photonics, Electronics, System design Motivating example: The design of a 100 Tb/s Internet router Challenging but not impossible (~100x current commercial systems) It identifies some interesting research problems
5
5 Arbitration 160Gb/s 40Gb/s Optical Switch Line termination IP packet processing Packet buffering Line termination IP packet processing Packet buffering 160- 320Gb/s 160- 320Gb/s Electronic Linecard #1 Electronic Linecard #625 Request Grant (100Tb/s = 625 * 160Gb/s) 100Tb/s optical router
6
6 Research Problems Linecard Memory bottleneck: Address lookup and packet buffering. Architecture Arbitration: Computation complexity. Switch Fabric Optics: Fabric scalability and speed, Electronics: Switch control and link electronics, Packaging: Three surface problem.
7
7 Packet Buffering Problem Packet buffers for a 40Gb/s router linecard Buffer Memory Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns 10Gbits Buffer Manager
8
8 Memory Technology Use SRAM? + Fast enough random access time, but - Too low density to store 10Gbits of data. Use DRAM? + High density means we can store data, but - Can’t meet random access time.
9
9 Can’t we just use lots of DRAMs in parallel? Buffer Memory Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns Buffer Manager Buffer Memory Buffer Memory Buffer Memory Buffer Memory Buffer Memory Buffer Memory Buffer Memory Read/write 320B every 32ns 40-79Bytes: 0-39……………280-319 320B
10
10 Works fine if there is only one FIFO Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns Buffer Manager 40-79Bytes: 0-39……………280-319 320B Buffer Memory 320B 40B 320B 40B 320B
11
11 In practice, buffer holds many FIFOs 40-79Bytes: 0-39……………280-319 320B 1 2 Q e.g. In an IP Router, Q might be 200. In an ATM switch, Q might be 10 6. Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns Buffer Manager 320B ?B 320B ?B How can we write multiple packets into different queues?
12
12 Arriving Packets R Arbiter or Scheduler Requests Departing Packets R 12 1 Q 2 1 2 34 34 5 123456 Small head SRAM cache for FIFO heads SRAM Hybrid Memory Hierarchy Large DRAM memory holds the body of FIFOs 5768109 798 11 12141315 5052515354 868887899190 8284838586 929493 95 68791110 1 Q 2 Writing b bytes Reading b bytes cache for FIFO tails 55 56 9697 87 88 57585960 899091 1 Q 2 Small tail SRAM DRAM
13
13 160Gb/s Linecard: Packet Buffering Solution Hybrid solution uses on-chip SRAM and off-chip DRAM. Identified optimal algorithms that minimize size of SRAM (12 Mbits). Precisely emulates behavior of 40 Gbit SRAM. DRAM 160 Gb/s Queue Manager klamath.stanford.edu/~nickm/papers/ieeehpsr2001.pdf SRAM
14
14 Research Problems Linecard Memory bottleneck: Address lookup and packet buffering. Architecture Arbitration: Computation complexity. Switch Fabric Optics: Fabric scalability and speed, Electronics: Switch control and link electronics, Packaging: Three surface problem.
15
15 Arbitration 160Gb/s 40Gb/s Optical Switch Line termination IP packet processing Packet buffering Line termination IP packet processing Packet buffering 160- 320Gb/s 160- 320Gb/s Electronic Linecard #1 Electronic Linecard #625 Request Grant (100Tb/s = 625 * 160Gb/s) 100Tb/s optical router
16
16 The Arbitration Problem A packet switch fabric is reconfigured for every packet transfer. At 160Gb/s, a new IP packet can arrive every 2ns. The configuration is picked to maximize throughput and not waste capacity. Known algorithms are too slow.
17
17 Cyclic Shift? 1 N 1 N Uniform Bernoulli iid traffic: 100% throughput Problem: real traffic is non-uniform
18
18 Two-Stage Switch External Outputs Internal Inputs 1 N External Inputs Load-balancing cyclic shift Switching cyclic shift 1 N 1 N 1 1 2 2 100% throughput for broad range of traffic types (C.S. Chang et al., 2001)
19
19 External Outputs Internal Inputs 1 N External Inputs Cyclic Shift 1 N 1 N 1 1 2 2 Problem: mis-sequencing
20
20 Preventing Mis-sequencing 1 N 1 N 1 N The Full Frames First algorithm: Keeps packets ordered and Guarantees a delay bound within the optimum Infocom’02: klamath.stanford.edu/~nickm/papers/infocom02_two_stage.pdf Small Coordination Buffers & ‘FFF’ Algorithm Large Congestion Buffers Cyclic Shift
21
21 Conclusions Packet Buffering Emulation of SRAM speed with DRAM density Packet buffer for a 160 Gb/s linecard is feasible Arbitration Developed Full Frames First Algorithm 100% throughput without scheduling
22
22 Two-Stage Switch External Outputs Internal Inputs 1 N External Inputs 1 N 1 N 1 (t) 2 (t) b(t) q(t) a(t) Traffic rate: First cyclic shift: Long-term service opportunities exceed arrivals:
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.