Presentation is loading. Please wait.

Presentation is loading. Please wait.

Digital Video Cluster Simulation Martin Milkovits CS699 – Professional Seminar April 26, 2005.

Similar presentations


Presentation on theme: "Digital Video Cluster Simulation Martin Milkovits CS699 – Professional Seminar April 26, 2005."— Presentation transcript:

1 Digital Video Cluster Simulation Martin Milkovits CS699 – Professional Seminar April 26, 2005

2 Goal of Simulation Build an accurate performance model of the interconnecting fabrics in a Digital Video cluster Build an accurate performance model of the interconnecting fabrics in a Digital Video cluster Assumptions Assumptions RAID Controller would follow a triangular distribution of I/O interarrival times RAID Controller would follow a triangular distribution of I/O interarrival times Gigabit Ethernet IP edge card would not impress any backpressure on the I/Os Gigabit Ethernet IP edge card would not impress any backpressure on the I/Os

3 Fabrics Simulated Table 1: Connection Technologies FabricType Per Link /Bus Actual Bandwidth (Gbps) Hardware Device BuffersPorts PCI 2.2 66MHz/ 64bit Parallel – bridged 3.934n/an/an/a StarFabric (SF) Full Duplex Serial 1.77 StarGen 2010 Bridge Per SF Port and PCI 2 – StarFabric 1 – 64/66 PCI StarGen 1010 Switch Per SF Port 6 – StarFabric InfiniBand (IB) Full Duplex Serial 2.0 Mellanox 21108 Bridge / Switch Per IB port and PCI 8 – 1X InfiniBand 1 – 64/66 PCI

4 Digital Video Cluster

5 Digital Video Node

6 Modules, Connections and Messages Messages represent data packets AND are used to control the model Messages represent data packets AND are used to control the model For data packets – have a non-zero length parameter For data packets – have a non-zero length parameter Contain routing and source information Contain routing and source information Modules handle message processing and routing Modules handle message processing and routing By and large represent hardware in the system By and large represent hardware in the system PCI Bus module – not actual hardware, but necessary to simulate a bus architecture PCI Bus module – not actual hardware, but necessary to simulate a bus architecture Connections allow messages to flow between modules Connections allow messages to flow between modules represent links/busses represent links/busses Independent connections for data vs. control messages Independent connections for data vs. control messages May be configured with a data rate value to simulate transmission delay May be configured with a data rate value to simulate transmission delay

7 Managing Buffer/Bus access Before transferring a data message (RWM) Need to gain access to transfer link/bus and destination buffer

8 PCI Bus Challenges Maintain Bus fairness Maintain Bus fairness Allow multiple PCI bus masters to interleave transactions (account for retry overhead) Allow multiple PCI bus masters to interleave transactions (account for retry overhead) Allow bursting if only one master Allow bursting if only one master

9 PCI Bus Module Components Queue – pending RWM’s Queue – pending RWM’s pciBus[maxDevices] array – utilization key pciBus[maxDevices] array – utilization key reqArray[maxDevices] – pending rqst messages reqArray[maxDevices] – pending rqst messages Work area – manages RWM actually being transferred by the PCI bus Work area – manages RWM actually being transferred by the PCI bus 3 Message types to handle 3 Message types to handle rqst messages from PCI bus masters rqst messages from PCI bus masters RMW messages RMW messages qCheck self-messages qCheck self-messages

10 Handling rqst and RWM messages When RWM finally hits the work area When RWM finally hits the work area Set RMW.transfer value = length of message (1024) Set RMW.transfer value = length of message (1024) Schedule qCheck self-message to fire in 240ns (time to transfer 128bits) Schedule qCheck self-message to fire in 240ns (time to transfer 128bits)

11 Handling qCheck Messages

12 Determining Max Bandwidth

13 Simulation Ramp-up

14 105Second @ 120MBps Results NodeRAID Sample Mean 90 % Confidence Interval 00119.9970.002 1119.9990.002 10120.0010.004 1119.9980.002 20120.0000.005 1120.0020.004 30120.0050.005 1120.0070.009 40119.9990.002 1120.0000.002 50119.9980.004 1120.0020.002 60119.9990.005 1120.0010.004

15 Contention / Utilization / Capacity

16 Learning Experiences PCI Contention PCI Contention First as a link like any other maintained by the StarGen chip First as a link like any other maintained by the StarGen chip Buffer contention and access Buffer contention and access Originally used retry loops – like actual system - way too much processing time! Originally used retry loops – like actual system - way too much processing time! Retry messages that are returned are a natural design given the language of messages and connections. Retry messages that are returned are a natural design given the language of messages and connections.

17 Conclusion / Future Work Simulation performed within 7% of actual system performance Simulation performed within 7% of actual system performance PCI bus between IB and StarGen potential hotspot PCI bus between IB and StarGen potential hotspot Complete more iterations with minor system modifications (dualDMA, scheduling) Complete more iterations with minor system modifications (dualDMA, scheduling) Submitted paper to the Winter Simulation Conference Submitted paper to the Winter Simulation Conference


Download ppt "Digital Video Cluster Simulation Martin Milkovits CS699 – Professional Seminar April 26, 2005."

Similar presentations


Ads by Google