SPEC MPI2007 Benchmarks for HPC Systems Ron Lieberman Chair, SPEC HPG HP-MPI Performance Hewlett-Packard Company Dr. Tom Elken Manager, Performance Engineering.

Slides:



Advertisements
Similar presentations
SPEC High Performance Group (An Introduction)
Advertisements

Doc.: IEEE /0006r0 Submission March 2005 Steve Shellhammer, Intel CorporationSlide 1 What is a CA document? Notice: This document has been prepared.
Tom Sugden EPCC OGSA-DAI Future Directions OGSA-DAI User's Forum GridWorld 2006, Washington DC 14 September 2006.
Cultural Heritage in REGional NETworks REGNET Project Meeting Content Group Part 1: Usability Testing.
Technology Analysis LINUX Alper Alansal Brian Blumberg Ramank Bharti Taihoon Lee.
MPI Message Queue Debugging Interface Chris Gottbrath Director, Product Management.
SPEC ENV2002: Environmental Simulation Benchmarks Wesley Jones
SPEC OMP Benchmark Suite H. Saito, G. Gaertner, W. Jones, R. Eigenmann, H. Iwashita, R. Lieberman, M. van Waveren, and B. Whitney SPEC High-Performance.
HPC Benchmarking, Rudi Eigenmann, 2006 SPEC Benchmark Workshop this is important for machine procurements and for understanding where HPC technology is.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller SPEC Benchmarks for Large Systems Matthias Mueller, Kumaran Kalyanasundaram, G. Gaertner, W. Jones,
SPEC HPG Benchmarks for HPC Systems Kumaran Kalyanasundaram for SPEC High-Performance Group Kumaran Kalyanasundaram, PhD Chair, SPEC HPG Manager, SGI Performace.
Höchstleistungsrechenzentrum Stuttgart Matthias M üller SPEC Benchmarks at HLRS Matthias Mueller High Performance Computing Center Stuttgart
NGS computation services: API's,
Configuration management
The following 10 questions test your knowledge of desired configuration management in Configuration Manager Configuration Manager Desired Configuration.
© 2003 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Performance Measurements of a User-Space.
INTRODUCTION TO SIMULATION WITH OMNET++ José Daniel García Sánchez ARCOS Group – University Carlos III of Madrid.
Full-System Timing-First Simulation Carl J. Mauer Mark D. Hill and David A. Wood Computer Sciences Department University of Wisconsin—Madison.
Performance, Area and Bandwidth Implications on Large-Scale CMP Cache Design Li Zhao, Ravi Iyer, Srihari Makineni, Jaideep Moses, Ramesh Illikkal, Don.
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
1 Jacob Thomas Basu Vaidyanathan Bret Olszewski Session II April 2014 POWER8 Benchmark and Performance.
The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
Introductions to Parallel Programming Using OpenMP
Transforming Business with Advanced Analytics: Introducing the New Intel® Xeon® Processor E7 v2 Family Seetha Rama Krishna Director, APAC HPC Solutions.
Prof. Srinidhi Varadarajan Director Center for High-End Computing Systems.
Network Redesign and Palette 2.0. The Mission of GCIS* Provide all of our users optimal access to GCC’s technology resources. *(GCC Information Services:
Job Submission on WestGrid Feb on Access Grid.
Introduction to UNIX Acknowledgement:Thanks to Dr Andrew Horner for the original version of this set of slides. All trademarks are the properties of their.
A Comparative Study of Network Protocols & Interconnect for Cluster Computing Performance Evaluation of Fast Ethernet, Gigabit Ethernet and Myrinet.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Introduction to UNIX Acknowledgement:Thanks to Dr Andrew Horner for the original version of this set of slides. All trademarks are the properties of their.
PROJECT PRESENTATION Prof: Daniel Amyot Presented By… ANVESH ALUWALA GURPREET SINGH DHADDA Evaluation of Load Testing Tools WebLOAD Professional Vs NeoLoad.
© 2005, it - instituto de telecomunicações. Todos os direitos reservados. System Level Resource Discovery and Management for Multi Core Environment Javad.
Chapter 10 – UNIX. History In late 1960s, two employees of Bell Labs (Ken Thompson & Dennis Ritchie) designed a new operating system to overcome the constraints.
Computer Performance Computer Engineering Department.
United Nations Economic Commission for Europe Statistical Division Seasonal Adjustment Process with Demetra+ Anu Peltola Economic Statistics Section, UNECE.
Copyright © 2002, Intel Corporation. All rights reserved. *Other brands and names are the property of their respective owners
3 rd Party Software Gail Alverson August 5, 2005.
Open MPI Project State of the Union - April 2007 Jeff Squyres Cisco, Inc.
Submission Title: Conformance, Interoperability & Certification of ZigBee Devices Date Submitted: 2 nd June 2003 Source: Bhupender Virk Company: ZigBee.
® IBM Software © IBM Corporation IBM Internal Use Only--Not to be shared outside the company until July 25, 2006 Processor Value Unit Licensing for Middleware.
Page 1 TOOLKIT / HDF-EOS Status and Development Abe Taaheri, Raytheon IIS Aura DSWG meeting October 2007.
IODE Ocean Data Portal - ODP  The objective of the IODE Ocean Data Portal (ODP) is to facilitate and promote the exchange and dissemination of marine.
© 2009 IBM Corporation Parallel Programming with X10/APGAS IBM UPC and X10 teams  Through languages –Asynchronous Co-Array Fortran –extension of CAF with.
Itanium 2 Impact Software / Systems MSC.Software Jay Clark Director, Business Development High Performance Computing
12/5/2015 Communication on Progress Elena Panova UN Global Compact Network Bulgaria.
Capacity and Capability Computing using Legion Anand Natrajan ( ) The Legion Project, University of Virginia (
Threaded Programming Lecture 2: Introduction to OpenMP.
Computing R&D and Milestones LHCb Plenary June 18th, 1998 These slides are on WWW at:
25 April Unified Cryptologic Architecture: A Framework for a Service Based Architecture Unified Cryptologic Architecture: A Framework for a Service.
edit type on title master Fortran ISV Release I to L LINPACK TOP500 Technical Systems Division * Scalable Computing Lab 2 Hsin-Ying Lin
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
Benchmark 2 progress report and future directions Sergey Lyskov
Doc.: IEEE /0287r2 Submission January 2006 Adrian Stephens, Intel Corporation TGn Editor Candidate – Adrian Stephens Notice: This document has.
CEG 2400 FALL 2012 Linux/UNIX Network Operating Systems.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
WS ►I Promoting Web services interoperability across platforms, applications and programming languages October, 2002.
The Multikernel: A New OS Architecture for Scalable Multicore Systems
High Performance Computing on an IBM Cell Processor --- Bioinformatics
Structural Simulation Toolkit / Gem5 Integration
TGn Editor Candidate – Adrian Stephens
Objective Understand the concepts of modern operating systems by investigating the most popular operating system in the current and future market Provide.
Chapter 2: The Linux System Part 1
CMSC 611: Advanced Computer Architecture
Web Services Interoperability Organization
Future Requirements of WIS Centres
CMSC 611: Advanced Computer Architecture
Section 1: Linux Basics and SLES9 Installation
Objective Understand the concepts of modern operating systems by investigating the most popular operating system in the current and future market Provide.
Presentation transcript:

SPEC MPI2007 Benchmarks for HPC Systems Ron Lieberman Chair, SPEC HPG HP-MPI Performance Hewlett-Packard Company Dr. Tom Elken Manager, Performance Engineering QLogic Corporation Dr William Brantley Manager HPC Performance AMD Dr. Matthias S. Mueller Vice Chair, SPEC HPG Deputy Director, CTO Center for Information Services and High Performance Computing (ZIH) Dresden University of Technology Dr Matthijs van Waveren Secretary, SPEC HPG Fujitsu Systems Europe Ltd

CAUTIONS The information contained within this presentation is a forward looking statement. Additionally, any slides with performance data are to be considered ESTIMATES and are labeled as such.

An application benchmark suite that measures CPU, memory, interconnect, compiler, MPI, and file system performance. Search program ended 3/31/06 Candidate codes in the areas of Comp. Chemistry, Weather, HE Physics, Oceanography, CFD, etc. SPEC MPI2007

CPU2006/MPI2007 Similarities Same tools used to run the benchmarks Similar run and reporting rules Uses geometric mean to calculate overall performance relative to a baseline system Similar output format

Comparison of benchmark characteristics

SPEC MPI2007 Development Participating Members AMD, Fujitsu, HP, IBM, INTEL, QLogic (PathScale), SGI, SUN, University of Dresden Current release targeted for March-June 2007 ISC07 in Dresden June 2007 most likely release. We are always looking for new members to help develop benchmarks

SPEC MPI2007 Benchmark Goals Runs on Clusters or SMPs Validates for correctness and measures performance Supports 32-bit or 64-bit OS/ABI. Consists of applications drawn from National Labs and University research centers Supports a broad range of MPI implementations and Operating systems including Windows, Linux, Proprietary Unix

SPEC MPI2007 Benchmark Goals Scales up and scales out Has a runtime of ~1 hour per benchmark test at 16 ranks using GigE with 1 GB memory footprint per rank Is extensible to future large and extreme data sets planned to cover larger number of ranks.

SPEC MPI Current Status Runs on range of architectures Opteron, Xeon, Itanium2, PA-Risc, Power5, Sparc, Ported to variety of operating systems Linux (RH/XC, SuSE, FC), Windows CCS, HPUX, Solaris, AIX Broad range of MPIs evaluated HP-MPI, MPICH, MPICH2, Open MPI, IBM-MPI, Intel MPI, MPICH-GM, MVAPICH, Fujitsu MPI, InfiniPath MPI, SGI MPT

SPEC MPI Current Status Scalable from 16 to 128 ranks (processes) for medium data set. 16 of 18 benchmarks run at 512 ranks. Runtime of 1 hour per benchmark test at 16 ranks using GigE on an unspecified reference cluster. Memory footprint should be < 1GB per rank at 16 ranks. Exhaustively tested for each rank count > , 160, 180, 200, 225, 256, 512

MPI2007 Performance Dimensions Scale out/upClusters, SMPs, Fatnode clusters Launch strategies affinity, process placement MPI Distributions open source, industrial, collective algorithms Operating systems distributions, kernel revisions, tunables Interconnectshardware, protocol, drivers, multi-rail HardwareCPU, memory, motherboards Compilersoptimization, correctness File SystemsDisks, Software, Network

SPEC MPI Likely Uses Customer RFPs Marketing messages as it relates to publication on SPEC HPG Web site. Academic Research Product Qualification and Performance evaluation Evaluate new releases, interconnects, OSs…

SPEC MPI2007 Benchmark Characteristics CategoryLanguage#LOC Physics C18K CFD FORTRAN11K CEM FORTRAN22K CFD FORTRAN & C45K Bioinformatics C11K Bioinformatics C++1,421K Oceanography FORTRAN71K Ray Tracing C16K Molecular Dynamics C++58K Weather Forecasting FORTRAN & C218K FEM (HT) FORTRAN & C31K Hydrodynamics FORTRAN7K Chemistry FORTRAN & C93K Hydrodynamics FORTRAN45K Abinitio C260K Ocean & Atm. FORTRAN & C41K Gravitation C24K CFD FORTRAN6K

SPEC MPI2007 (32 ranks) Characteristics -- ESTIMATES

MPI2007 Benchmark Message call counts

SPEC MPI2007 Fair Use Policy SPEC/HPG Fair Use Rule Consistency and fairness are guiding principles for SPEC. To assure these principles are sustained, the following guidelines have been created with the intent that they serve as specific guidance for any organization (or individual) who chooses to make public comparisons using SPEC benchmark results. When any organization, including vendor or research oriented organizations, or any individual makes public claims using SPEC benchmark results, SPEC requires that the following guidelines be observed: [1] Reference is made to the SPEC trademark. Such reference may be included in a notes section with other trademark references (see for all SPEC trademarks and service marks). [2] The SPEC web site ( or a suitable sub page is noted as the source for more information. [3] If any public claims or competitive comparisons are made, the results stated or compared must be compliant with that benchmark's run and reporting rules and must cite the following: SPEC metric, CPU description (number of chips and cores), and number of OpenMP threads and/or MPI ranks. [4] If competitive comparisons are made the following rules apply:a. the basis for comparison must be stated, b. the source of the competitive data must be stated, c. the date competitive data was retrieved must be stated, d. all data used in comparisons must be publicly available (from SPEC or elsewhere) e. the benchmark must be currently accepting new submissions if previously unpublished results are used in the comparison. [5] Comparisons with or between non-compliant test results can only be made within academic or research documents or presentations where the deviations from the rules for any non-compliant results have been disclosed. A compliant test result is a test result that has followed the run rules, and has been submitted and approved by SPEC. SPEC HPG makes recommendations for the academic or research use of benchmark results in the document, "Guidelines for the Use of SPEC HPG Benchmarks in Research Publications."."Guidelines for the Use of SPEC HPG Benchmarks in Research Publications."

SPEC MPI2007 RunRules This document specifies how the benchmarks in the MPI2007 suites are to be run for measuring and publicly reporting performance results, to ensure that results generated with the suites are meaningful, comparable to other generated results, and reproducible (with documentation covering factors pertinent to reproducing the results). Per the SPEC license agreement, all results publicly disclosed must adhere to the SPEC Run and Reporting Rules, or be clearly marked as estimates.

Active members of SPEC HPG who make things happen with their dedication and passion AMD, Fujitsu, HP, IBM, Intel, QLogic, SGI, SUN, University of Dresden SPEC OSG for allowing us to leverage CPU2006 benchmarks and tools Have I mentioned we are always looking for new members to help develop benchmarks? Acknowledgements