Ph. D Oral Examination Load-Balancing and Parallelism for the Internet Stanford University Ph.D. Oral Examination Tuesday, Feb 18 th 2003 Sundar Iyer

Slides:



Advertisements
Similar presentations
Nick McKeown Spring 2012 Lecture 4 Parallelizing an OQ Switch EE384x Packet Switch Architectures.
Advertisements

Sundar Iyer Winter 2012 Lecture 8a Packet Buffers with Latency EE384 Packet Switch Architectures.
Fast Buffer Memory with Deterministic Packet Departures Mayank Kabra, Siddhartha Saha, Bill Lin University of California, San Diego.
Optimal-Complexity Optical Router Hadas Kogan, Isaac Keslassy Technion (Israel)
1 Statistical Analysis of Packet Buffer Architectures Gireesh Shrimali, Isaac Keslassy, Nick McKeown
Shivkumar Kalyanaraman Rensselaer Polytechnic Institute 1 High Speed Router Design Shivkumar Kalyanaraman Rensselaer Polytechnic Institute
Router Architecture : Building high-performance routers Ian Pratt
Nick McKeown CS244 Lecture 6 Packet Switches. What you said The very premise of the paper was a bit of an eye- opener for me, for previously I had never.
Routers with a Single Stage of Buffering Sundar Iyer, Rui Zhang, Nick McKeown High Performance Networking Group, Stanford University,
Isaac Keslassy, Shang-Tse (Da) Chuang, Nick McKeown Stanford University The Load-Balanced Router.
Algorithm Orals Algorithm Qualifying Examination Orals Achieving 100% Throughput in IQ/CIOQ Switches using Maximum Size and Maximal Matching Algorithms.
Making Parallel Packet Switches Practical Sundar Iyer, Nick McKeown Departments of Electrical Engineering & Computer Science,
Analysis of a Statistics Counter Architecture Devavrat Shah, Sundar Iyer, Balaji Prabhakar & Nick McKeown (devavrat, sundaes, balaji,
1 Input Queued Switches: Cell Switching vs. Packet Switching Abtin Keshavarzian Joint work with Yashar Ganjali, Devavrat Shah Stanford University.
The Concurrent Matching Switch Architecture Bill Lin (University of California, San Diego) Isaac Keslassy (Technion, Israel)
Analyzing Single Buffered Routers Sundar Iyer, Rui Zhang, Nick McKeown (sundaes, rzhang, High Performance Networking Group Departments.
Analysis of a Packet Switch with Memories Running Slower than the Line Rate Sundar Iyer, Amr Awadallah, Nick McKeown Departments.
1 Architectural Results in the Optical Router Project Da Chuang, Isaac Keslassy, Nick McKeown High Performance Networking Group
1 OR Project Group II: Packet Buffer Proposal Da Chuang, Isaac Keslassy, Sundar Iyer, Greg Watson, Nick McKeown, Mark Horowitz
Using Load-Balancing To Build High-Performance Routers Isaac Keslassy, Shang-Tse (Da) Chuang, Nick McKeown Stanford University.
048866: Packet Switch Architectures Dr. Isaac Keslassy Electrical Engineering, Technion Scaling.
1 Internet Routers Stochastics Network Seminar February 22 nd 2002 Nick McKeown Professor of Electrical Engineering and Computer Science, Stanford University.
EE 122: Router Design Kevin Lai September 25, 2002.
Nick McKeown 1 Memory for High Performance Internet Routers Micron February 12 th 2003 Nick McKeown Professor of Electrical Engineering and Computer Science,
1 EE384Y: Packet Switch Architectures Part II Load-balanced Switches Nick McKeown Professor of Electrical Engineering and Computer Science, Stanford University.
Maximum Size Matchings & Input Queued Switches Sundar Iyer, Nick McKeown High Performance Networking Group, Stanford University,
1 Trend in the design and analysis of Internet Routers University of Pennsylvania March 17 th 2003 Nick McKeown Professor of Electrical Engineering and.
1 Achieving 100% throughput Where we are in the course… 1. Switch model 2. Uniform traffic  Technique: Uniform schedule (easy) 3. Non-uniform traffic,
Analysis of a Memory Architecture for Fast Packet Buffers Sundar Iyer, Ramana Rao Kompella & Nick McKeown (sundaes,ramana, Departments.
Surprise Quiz EE384Z: McKeown, Prabhakar ”Your Worst Nightmares in Packet Switching Architectures”, 3 units [Total time = 15 mins, Marks: 15, Credit is.
1 Growth in Router Capacity IPAM, Lake Arrowhead October 2003 Nick McKeown Professor of Electrical Engineering and Computer Science, Stanford University.
Pipelined Two Step Iterative Matching Algorithms for CIOQ Crossbar Switches Deng Pan and Yuanyuan Yang State University of New York, Stony Brook.
1 IP routers with memory that runs slower than the line rate Nick McKeown Assistant Professor of Electrical Engineering and Computer Science, Stanford.
Computer Networks Switching Professor Hui Zhang
Load Balanced Birkhoff-von Neumann Switches
Belgrade University Aleksandra Smiljanić: High-Capacity Switching Switches with Input Buffers (Cisco)
Summary of switching theory Balaji Prabhakar Stanford University.
Designing Packet Buffers for Internet Routers Friday, October 23, 2015 Nick McKeown Professor of Electrical Engineering and Computer Science, Stanford.
Designing Packet Buffers for Router Linecards Sundar Iyer, Ramana Kompella, Nick McKeown Reviewed by: Sarang Dharmapurikar.
Winter 2006EE384x1 EE384x: Packet Switch Architectures I Parallel Packet Buffers Nick McKeown Professor of Electrical Engineering and Computer Science,
Routers. These high-end, carrier-grade 7600 models process up to 30 million packets per second (pps).
Applied research laboratory 1 Scaling Internet Routers Using Optics Isaac Keslassy, et al. Proceedings of SIGCOMM Slides:
Packet Forwarding. A router has several input/output lines. From an input line, it receives a packet. It will check the header of the packet to determine.
Nick McKeown1 Building Fast Packet Buffers From Slow Memory CIS Roundtable May 2002 Nick McKeown Professor of Electrical Engineering and Computer Science,
1 Performance Guarantees for Internet Routers ISL Affiliates Meeting April 4 th 2002 Nick McKeown Professor of Electrical Engineering and Computer Science,
Stress Resistant Scheduling Algorithms for CIOQ Switches Prashanth Pappu Applied Research Laboratory Washington University in St Louis “Stress Resistant.
An Introduction to Packet Switching Nick McKeown Assistant Professor of Electrical Engineering and Computer Science, Stanford University
Nick McKeown Spring 2012 Lecture 2,3 Output Queueing EE384x Packet Switch Architectures.
Winter 2006EE384x1 EE384x: Packet Switch Architectures I a) Delay Guarantees with Parallel Shared Memory b) Summary of Deterministic Analysis Nick McKeown.
Belgrade University Aleksandra Smiljanić: High-Capacity Switching Switches with Input Buffers (Cisco)
T. S. Eugene Ngeugeneng at cs.rice.edu Rice University1 COMP/ELEC 429 Introduction to Computer Networks Lecture 18: Quality of Service Slides used with.
Techniques for Fast Packet Buffers Sundar Iyer, Ramana Rao, Nick McKeown (sundaes,ramana, Departments of Electrical Engineering & Computer.
Buffered Crossbars With Performance Guarantees Shang-Tse (Da) Chuang Cisco Systems EE384Y Thursday, April 27, 2006.
1 A quick tutorial on IP Router design Optics and Routing Seminar October 10 th, 2000 Nick McKeown
1 How scalable is the capacity of (electronic) IP routers? Nick McKeown Professor of Electrical Engineering and Computer Science, Stanford University
Techniques for Fast Packet Buffers Sundar Iyer, Ramana Rao, Nick McKeown (sundaes,ramana, Departments of Electrical Engineering & Computer.
Packet Switch Architectures The following are (sometimes modified and rearranged slides) from an ACM Sigcomm 99 Tutorial by Nick McKeown and Balaji Prabhakar,
Block-Based Packet Buffer with Deterministic Packet Departures Hao Wang and Bill Lin University of California, San Diego HSPR 2010, Dallas.
The Fork-Join Router Nick McKeown Assistant Professor of Electrical Engineering and Computer Science, Stanford University
Techniques for Fast Packet Buffers Sundar Iyer, Nick McKeown Departments of Electrical Engineering & Computer Science, Stanford.
Network layer (addendum) Slides adapted from material by Nick McKeown and Kevin Lai.
1 Building big router from lots of little routers Nick McKeown Assistant Professor of Electrical Engineering and Computer Science, Stanford University.
scheduling for local-area networks”
Weren’t routers supposed
Packet Forwarding.
Addressing: Router Design
Parallelism in Network Systems Joint work with Sundar Iyer
Write about the funding Sundar Iyer, Amr Awadallah, Nick McKeown
Techniques and problems for
Techniques for Fast Packet Buffers
Presentation transcript:

Ph. D Oral Examination Load-Balancing and Parallelism for the Internet Stanford University Ph.D. Oral Examination Tuesday, Feb 18 th 2003 Sundar Iyer Department of Computer Science Stanford University,

Ph. D Oral Examination2 Motivation  Build routers with performance guarantees  Guarantees are  Statistical (100% throughput)  Deterministic (work-conserving, delay guarantees)  Hard to build big routers with guarantees  This talk:  Deterministic guarantees  Use load-balancing and parallelism

Ph. D Oral Examination3 Contents 1. Parallel routers: work-conserving 2. Parallel routers: delay guarantees 3. Parallel packet buffers 4. Summary of contributions

Ph. D Oral Examination4 Output-Queued Router one output G/D/1  Work-conservation :  100% throughput, minimizes delay, delay guarantees possible  Problem : Memory bandwidth not scalable  Previous attempts: Very complicated or ad hoc  My approach: Deterministic parallel work-conserving output-queued router

Ph. D Oral Examination5 Parallel Output-Queued Router one output 1 k many outputs 1 k N N

Ph. D Oral Examination6 At most two memory operations per time slot: 1 write and 1 read Parallel Output-Queued Router (May not be work-conserving) 1 1 k=3 N=3 A C B 2 Time slot = 1 A5 A6 A7 A5 A6 A7 B5 B6 A8 B5 B6 A8 Time slot = 2 B6 B5 A8 C5 C6 Time slot = 3 Constant size packets

Ph. D Oral Examination7  Problem :  How can we design a parallel output-queued work- conserving router from slower parallel memories? Problem  Theorem 1: (sufficiency)  A parallel output-queued router is work- conserving with 3N –1 memories that can perform at most one memory operation per time slot

Ph. D Oral Examination8 Re-stating the Problem  There are K cages which can contain an infinite number of pigeons.  Assume that time is slotted, and in any one time slot  at most N pigeons can arrive and at most N can depart.  at most 1 pigeon can enter or leave a cage via a pigeon hole.  the time slot at which arriving pigeons will depart is known  For any router:  What is the minimum K, such that all N pigeons can be immediately placed in a cage when they arrive, and can depart at the right time?

Ph. D Oral Examination9  Only one packet can enter or leave a memory at time t Intuition for Theorem 1  Only one packet can enter a memory at time t Time = t DT=t+X DT=t  Only one packet can enter or leave a memory at any time Memory

Ph. D Oral Examination10 Proof of Theorem 1  When a packet arrives in a time slot it must choose a memory not chosen by 1. The N – 1 other packets that arrive at that timeslot. 2. The N other packets that depart at that timeslot. 3. The N - 1 other packets that can depart at the same time as this packet departs (in future).  Proof:  By the pigeon-hole principle, 3N –1 memories that can perform at most one memory operation per time slot are sufficient for the router to be work-conserving

Ph. D Oral Examination11 Memory The Parallel Shared Memory Router C A Departing Packets R R Arriving Packets A5 A4 B1 C1 A1 C3 A5 A4 From theorem 1, k = 7 memories don’t suffice.. but 8 memories do Memory 1 K=8 C3 At most one operation – a write or a read per time slot B B3 C1 A1 A3 B1

Ph. D Oral Examination12 Distributed Shared Memory Router The central memories are moved to distributed line cards and shared Memory and line cards can be added incrementally From theorem 1, 3N –1 memories which can perform one operation per time slot i.e. a total memory bandwidth of  3NR suffice for the router to be work-conserving Switch Fabric Line Card 1Line Card 2Line Card N R RR Memories

Ph. D Oral Examination13 Corollary 1  Problem:  What is the switch bandwidth for a work-conserving DSM router?  Corollary 1: (sufficiency)  A switch bandwidth of 4NR is sufficient for a distributed shared memory router to be work-conserving  Proof 1:  There are a maximum of 3 memory accesses and 1 port access

Ph. D Oral Examination14 Corollary 2  Problem:  What is the switching algorithm for a work-conserving DSM router? Bus: No algorithm needed, but impractical Crossbar: Algorithm needed because only permutations are allowed  Corollary 2: (existence)  An edge coloring algorithm can switch packets for a work-conserving distributed shared memory router  Proof :  Follows from König’s theorem - Any bipartite graph with maximum degree  has an edge coloring with  colors

Ph. D Oral Examination15 Summary - Routers which give 100% throughput 2NR 2NR/kNk None Maximal2NR6NR3R2N MWM NR2NR2RNCrossbarInput Queued None2NR 1BusC. Shared Mem. Switch Algorithm Switch BW Total Mem. BW Mem. BW per Mem 1 # Mem.Fabric NoneNRN(N+1)R(N+1)RNBusOutput-Queued P Shared Mem. C. Sets4NR2N(N+1)R2R(N+1)/kNkClosPPS - OQ C. Sets4NR 4RN C. Sets6NR3NR3RN Edge Color4NR3NR3RN Xbar C. Sets2NR3NR3NR/kkBus C. Sets4NR 4NR/kNk Clos Time Reserve * 3NR6NR3R2N Crossbar PPS – Shared Memory DSM (Juniper) CIOQ (Cisco) 1 Note that lower mem. bandwidth per memory implies higher random access time, which is better

Ph. D Oral Examination16 Contents 1. Parallel routers: work-conserving 2. Parallel routers: delay guarantees 3. Parallel packet buffers 4. Summary of contributions

Ph. D Oral Examination17 Delay Guarantees one output, many logical FIFO queues 1 m Weighted fair queueing sorts packets constrained traffic PIFO models  Weighted Fair Queueing  Weighted Round Robin  Strict priority etc. one output, single PIFO queue Push In First Out (PIFO) constrained traffic push

Ph. D Oral Examination18 Delay Guarantees  Problem :  How can we design a parallel output-queued router from slower parallel memories and give delay guarantees?  This is difficult because  The counting technique depends on being able to predict the departure time and schedule it.  The departure time of a cell is not fixed in policies such as strict priority, weighted fair queueing etc.

Ph. D Oral Examination19 Theorem 2  Theorem 2: (sufficiency)  A parallel output-queued router can give delay guarantees with 4N –2 memories that can perform at most one memory operation per time slot.

Ph. D Oral Examination20 Intuition for Theorem 2 N=3 port router Departure Order … … Departure Order N-1 packets before cell at time of insertion DT = 3 DT= 2DT= 1 … N-1 packets after cell at time of insertion PIFO: 2 windows of memories of size N-1 that can’t be used FIFO: Window of memories of size N-1 that can’t be used Departure Order

Ph. D Oral Examination21 Proof of Theorem 2 A packet cannot use the memories: 1.Used to write the N-1 arriving cells at t. 2.Used to read the N departing cells at t. Time = t DT=t DT=t+T Cell C Before C After C 3.Will be used to read the N-1 cells that depart before it. 4.Will be used to read the N-1 cells that depart after it.

Ph. D Oral Examination22 Summary- Routers which give delay guarantees Marriage2NR6NR3R2N -2NR2NR2NR/kNk - NR2NR2RNCrossbar Input Queued None2NR 1BusC. Shared Mem. Switch Algorithm Switch BW Total Memory BW Mem. BW per Mem. # Mem.Fabric NoneNRN(N+1)R(N+1)RNBusOutput-Queued P. Shared M C. Sets6NR3N(N+1)R3R(N+1)/kNkClosPPS - OQ C. Sets6NR 6RN C. Sets8NR4NR4RN Edge Color5NR4NR4RN Xbar C. Sets2NR4NR4NR/kkBus C. Sets6NR 6NR/kNk Clos Time Reserve 3NR6NR3R2N Crossbar PPS – Shared Memory DSM (Juniper) CIOQ (Cisco)

Ph. D Oral Examination23 Contents 1. Parallel routers: work-conserving 2. Parallel routers: delay guarantees 3. Parallel packet buffers 4. Summary of contributions

Ph. D Oral Examination24 Packet Buffering  Big: For TCP to work well, the buffers need to hold one RTT (about 0.25s) of data.  Fast: Clearly, the buffer needs to store (retrieve) packets as fast as they arrive (depart). Input or Output Line CardShared Memory Buffer Line rate, R Memory 1 N 1 N Scheduler Memory Scheduler

Ph. D Oral Examination25 An Example Packet buffers for a 40Gb/s line card Buffer Memory Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns 10Gbits Buffer Manager Problem is solved if a memory can be (random) accessed every 4 ns and store 10Gb of data Scheduler requests causes random access

Ph. D Oral Examination26 Available Memory Technology  Use SRAM? + Fast enough random access time, but - Too low density to store 10Gbits of data.  Use DRAM? + High density means we can store data, but - Can’t meet random access time.

Ph. D Oral Examination27 Problem: How can we design high speed packet buffers from commodity available memories? Problem

Ph. D Oral Examination28 Can’t we just use lots of DRAMs in parallel? Buffer Memory Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns Buffer Manager Buffer Memory Buffer Memory Buffer Memory Buffer Memory Buffer Memory Buffer Memory Buffer Memory 320B 40B 320B Scheduler Requests

Ph. D Oral Examination29 Works fine if there is only one FIFO queue Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns Buffer Manager (on chip SRAM) 320B 40B 320B 40B 320B 40B 320B Scheduler Requests Aggregate 320B for the queue in fast SRAM and read and write to all DRAMs in parallel

Ph. D Oral Examination30 In practice, buffer holds many FIFOs 320B 1 2 Q e.g.  In an IP Router, Q might be 200.  In an ATM switch, Q might be Write Rate, R One 40B packet every 8ns Read Rate, R One 40B packet every 8ns Buffer Manager (on chip SRAM) 320B ?B 320B ?B We don’t know which head of line packet the scheduler will request next? 40B 320B Scheduler Requests

Ph. D Oral Examination31 Problems  Problems 1. A 320B block will contain packets for different queues, which can’t be written to, or read from the same location. 2. Suppose we write packets for different queues to different locations. How do we know that the memory will be available for reading when we need to retrieve the packet?  Requirement 1. We will have to aggregate 320B for every queue and read and write to all DRAMs in parallel

Ph. D Oral Examination32 Buffer Manager Arriving Packets R Scheduler Requests Departing Packets R 12 1 Q Small head SRAM cache for FIFO heads (ASIC with on chip SRAM) Parallel Packet Buffer Hybrid Memory Hierarchy cache for FIFO tails Q 2 Small tail SRAM Large DRAM memory holds the body of FIFOs Q 2 Writing b bytes Reading b bytes DRAM b = degree of parallelism

Ph. D Oral Examination33  Problem:  What is the minimum size of the SRAM needed for the parallel buffer so that every packet is available immediately in the SRAM when requested?  Theorem 3: (Necessity)  An SRAM size of size Qw = Q(b – 1)(2 + lnQ) bytes is necessary. w is the size of each queue Re-stating the Problem

Ph. D Oral Examination34 Why do we need a large SRAM? Q = 4, w = 3, b = 3 t = 3 Bytes t = 2 Bytes t = 1 Bytes Replenish t = 4 Bytes t = 5 Bytes t = 6 Bytes Replenish At t = 6 if the scheduler requests a black packet it will not be in the SRAM t = 0 Bytes Q w

Ph. D Oral Examination35 Theorem 4  Theorem 4: (sufficiency)  An SRAM cache of size Qw = Qb(2 +ln Q) bytes is sufficient so that every packet is available immediately in the SRAM when requested  Discussion:  The algorithm replenishes the queue with the most urgent need of replenishment  It is almost optimal

Ph. D Oral Examination36 Intuition for Theorem 4  The maximum number of un-replenished requests for any i queues w i, is the solution of the difference equation -  with boundary conditions Examples: 1.40Gb/s line card, b =640, Q =128: SRAM = 560kBytes 2.160Gb/s line card, b =2560, Q =512: SRAM = 10MBytes

Ph. D Oral Examination37  Problem:  What is the minimum size of the SRAM so that every packet is available in the SRAM within a bounded pipeline latency when requested? Theorem 5  Theorem 5: (necessity and sufficiency)  An SRAM cache of size Qw = Q(b – 1) bytes is both necessary and sufficient if the pipeline latency is Q(b – 1) + 1 time slots.

Ph. D Oral Examination38 Intuition for Theorem 5  The algorithm replenishes the queue which is going to be in the most urgent need of replenishment  If we use a lookahead buffer to know the requests “in advance”, we can identify the queue which will empty earliest  This increases the pipeline latency from when a request is made until the byte is available. Example: 1.160Gb/s line card, b =2560, Q =128: SRAM = 160kBytes, latency is 8  s.

Ph. D Oral Examination39 Theorem 6  Problem:  What is the minimum size of the SRAM needed so that every packet is available in the SRAM within a bounded pipeline latency x in the range (0, Q(b – 1) + 1) when requested?  Theorem 6: (sufficiency)  An SRAM cache of size Qw  Qb(2 +ln Qb/x) bytes is sufficient.

Ph. D Oral Examination Pipeline Latency, x Queue Length, w Queue Length for Zero Latency Queue Length for Maximum Latency Discussion of Theorem 6 Q=1000, b = 10

Ph. D Oral Examination41 Contents 1. Parallel routers: work-conserving 2. Parallel routers: delay guarantees 3. Parallel packet buffers 4. Summary of contributions

Ph. D Oral Examination42 Summary of Contributions Covered in this talk 1. Parallel routers: S. Iyer, R. Zhang, N. McKeown, “Routers with a Single Stage of Buffering", Proceedings of ACM SIGCOMM, Pittsburgh, Pennsylvania, Sep Also in Computer Communication Review, vol. 32, no. 4, Oct Parallel packet buffers: S. Iyer, R. R. Kompella, N. McKeown, “Designing Packet Buffers for Router Line Cards", Under review in IEEE/ACM Transactions on Networking, Oct S. Iyer, R. R. Kompella, N. McKeown, “Analysis of a Memory Architecture for Fast Packet Buffers", IEEE - High Performance Switching and Routing, pp , Dallas, May A preliminary version appeared in IEEE GBN, Alaska, April 2001.

Ph. D Oral Examination43 Summary of Contributions Not covered in this talk 3. Other switches (load-balancing): S. Iyer, N. McKeown, “Analysis of the Parallel Packet Switch Architecture", to appear in IEEE/ACM Transactions on Networking, Apr S. Iyer, A. Awadallah, N. McKeown, “Analysis of a Packet Switch with Memories Running Slower than the Line Rate", Proceedings of IEEE INFOCOM, pp , Tel Aviv, March S. Iyer, N. McKeown, “Making Parallel Packet Switches Practical", Proceedings of IEEE INFOCOM, Alaska,, vol. 3, pp , April S. Iyer, N. McKeown, “Using Constraint Sets to Achieve Delay Bounds in CIOQ Switches", to appear in IEEE Communication Letters, Parallel packet switch (multicast): S. Iyer, N. McKeown, “On the Speedup Required for a Multicast Parallel Packet Switch", IEEE Communication Letters, vol. 5, no. 6, pp , June 2001.

Ph. D Oral Examination44 Summary of Contributions Not covered in this talk 5. Parallel architecture (statistics counters): D. Shah, S. Iyer, B. Prabhakar, N. McKeown, “Maintaining Statistics Counters in Router Line Cards", IEEE Micro, pp , Jan-Feb, Also appeared as “Analysis of a Statistics Counter Architecture" in IEEE Hot Interconnects, Stanford University, Aug Parallel architecture (packet state): S. Iyer, N. McKeown, “Maintaining State in Router Line Cards“. In preparation for IEEE Communication Letters. 7. Parallel packet classification: S. Iyer, R. R. Kompella, A. Shelat, “ClassiPI: An Architecture for Fast and Flexible Packet Classification", IEEE NETWORK, Special Issue on Fast IP Packet Forwarding and Classification for Next Generation Internet Services, Mar-Apr S. Iyer, A. Desai, A. Tambe, A. Shelat, “ClassiPI: A Classifier for Next Generation Policy Based Engines", IEEE Hot Chips, Stanford University, Aug 2000.

Ph. D Oral Examination45 Summary of Contributions Not covered in this talk 8. Parallel switching algorithms (Buffered crossbars): Jointly with Da Chuang, N. McKeown, “Attaining Statistical and Deterministic Switching Guarantees using Buffered Crossbars", In preparation for IEEE/ACM Transactions on Networking. Jointly with Da Chuang, N. McKeown, “Designing Practical Routers using Buffered Crossbars“, In preparation for IEEE/ACM Transactions on Networking. 9. Load-balancing (IP backbone): S. Iyer, S. Bhattacharyya, N. Taft, C. Diot, “An Approach to Alleviate Link Overload as Observed on an IP Backbone“, Proceedings of IEEE INFOCOM, San Francisco, March Greedy switching algorithms: S. Iyer, N. McKeown, “Maximum Size Matchings and Input Queued Switches", Proceedings of the 40th Annual Allerton Conference on Communication, Control, and Computing, Monticello, Illinois, Oct 2002.

Ph. D Oral Examination46 Acknowledgements 1. Nick 2. Balaji 3. Members of my committee – Sunil, Mark and Rajeev 4. Industry: Cisco - Flavio Sprint Labs – Christophe, Supratik, Nina and the IP group SwitchOn/PMC-Sierra - Ajit, Moti and the whole team 5. Research “co-conspirators” 6. Members of the research group 7. Members of network-group 8. Department staff 9. Friends, friends & friends from all over the globe … 10. The couch on the 5 th floor, the basement vending machine, … 11. Mom  I was told: “A good way to judge how well you are doing, is to measure the quality of people with whom you interact.  By that criteria, I could not have asked for more ….