Presentation on theme: "Farms/Clusters of the future Large Clusters O(1000), any existing examples ? YesSupercomputing, PC clusters LLNL, Los Alamos, Google No long term experience."— Presentation transcript:
Farms/Clusters of the future Large Clusters O(1000), any existing examples ? YesSupercomputing, PC clusters LLNL, Los Alamos, Google No long term experience (except google) ASCI Installation : Powerstation (xMW), Well organized recover management (see Eniac)
Farms/Clusters of the future Reliability By Martin H. Weik, 1961 Ordnance Ballistic Research Laboratories, Aberdeen Proving Ground, MD The ENIAC's first few years at the Aberdeen Proving Ground were difficult ones for the operating and maintenance crews. … The result was a huge preventive-maintenance and testing program, which, in the end, led to some major modifications of the system. Tubes were life-tested, and statistical data on the failures were compiled. This information led to many improvements in vacuum tubes themselves. Procurement of large quantities of improved, reliable tubes, however, became a difficult problem. Power-line fluctuations and power failures made continuous operation directly off transformer mains an impossibility. The substantial quantity of heat which had to be dissipated into the warm, humid Aberdeen atmosphere created a heat-removal problem of major proportions. Down times were long; error-free running periods were short.
Farms/Clusters of the future Avoid local disks ?
Future CPUs - Low Power (mobile ?) ?
Architecture Results Amanda Simulation Amanda Reconstruction Host: oceanide1, new desktop PC, 533 MHz FSB Pentium4, 2.40GHz, 512 KB Cache, 256 MB Memory u, s u, 0.140s Host: minerva, theory PC, 400 MHz FSB XEON-P4, 2.00GHz, 512 KB Cache, 2 GB Memory u, 4.620s u, 0.120s Host: euterpe, network services Pentium III (Tualatin), 1266MHz, 512 KB Cache, 1 GB Memory u, 9.370s u, 0.740s Host: ice53, farm node Pentium III (Coppermine), 800 MHz, 256 KB cache 512 MB u, s u, 0.310s
LQCD - Dirac Operator Benchmark (SSE) 16x16 3, single P4/XEON CPU Dirac operator Linear Algebra MFLOPS Intensive cache pre-fetchNo benefit from cache
Future Cluster/Farm Architectures - Blade Servers ? NEXCOM – Low voltage blade server 200 low voltage Intel XEON CPUs (1.6 GHz – 30W) in a 42U Rack Integrated Gbit Ethernet network Mellanox – Infiniband blade server Single XEON Blades connected via a 10 Gbit (4X) Infiniband network MEGWARE, NCSA, Ohio State University
VID PSI IMVP IV Hub Interface ICH4-M DDR memory AC’ ModemModem Mobile Optimized Processor Target Avg. Power < 1W Mobile Optimized Processor Target Avg. Power < 1W USB 2.0 APIC Enabled Integrated Graphics Integrated Graphics DVO (2 ports) LVDS TPV DVO Optimized Power Supply 2 ATA66/100 IDE Channels 6 USB (1.1/2.0) Ports Integrated LAN PCI 33 MHz Cardbus 400MHz Low Power Processor System Bus b802.11b Intel® Pentium® M Processor Intel® 855GM Intel® Centrino ™ Mobile Technology Intel® 855GM chipset Intel® Pro/Wireless Network Connection
Benchmark list hyper threading vs non hyper threading Architecture Results Amanda Simulation Amanda Reconstruction Theorie form3 Host: cube3, farm node, 400 MHz FSB XEON-P4, 2.4 GHz, 512 KB cache, 2 GB Memeory Non hyper-threading kernel u 1.480s u 0.320sTime = sec Generated terms = Host: cube8, farm node, 400 MHz FSB XEON-P4, 2.4 GHz, 512 KB cache, 2 GB Memeory Hyper-threading kernel u 1.530s u 0.430sTime = sec Generated terms =