Alternative Performance Metrics for Server RFPs Joe Temple Low Country North Shore Consulting

Slides:



Advertisements
Similar presentations
Introduction to Grid Application On-Boarding Nick Werstiuk
Advertisements

Applying Benchmark Data To A Model for Relative Server Capacity CMG 2013 Joseph Temple, LC-NS Consulting John J Thomas, IBM.
Hadi Goudarzi and Massoud Pedram
The Who, What, Why and How of High Performance Computing Applications in the Cloud Soheila Abrishami 1.
Erhan Erdinç Pehlivan Computer Architecture Support for Database Applications.
Proactive Prediction Models for Web Application Resource Provisioning in the Cloud _______________________________ Samuel A. Ajila & Bankole A. Akindele.
Capacity Planning and Predicting Growth for Vista Amy Edwards, Ezra Freeloe and George Hernandez University System of Georgia 2007.
Ó 1998 Menascé & Almeida. All Rights Reserved.1 Part IV Capacity Planning Methodology.
Performance Engineering Methodology Chapter 4. Performance Engineering Performance engineering analyzes the expected performance characteristics of a.
Copyright © 2013, Oracle and/or its affiliates. All rights reserved.
1 Part IV Capacity Planning Methodology © 1998 Menascé & Almeida. All Rights Reserved.
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
Performance Evaluation
Traffic Characterization Dr. Abdulaziz Almulhem. Almulhem©20012 Agenda Traffic characterization Switching techniques Internetworking, again.
1 CS 501 Spring 2005 CS 501: Software Engineering Lecture 22 Performance of Computer Systems.
Introduction to the new mainframe: Large-Scale Commercial Computing © Copyright IBM Corp., All rights reserved. Chapter 3: Scalability.
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Matei Ripeanu.
KPIs and Machine Comparisons Joe Temple Copyright 5/2014 t Low Country North Shore Consulting.
Towards Autonomic Hosting of Multi-tier Internet Services Swaminathan Sivasubramanian, Guillaume Pierre and Maarten van Steen Vrije Universiteit, Amsterdam,
Capacity Planning in SharePoint Capacity Planning Process of evaluating a technology … Deciding … Hardware … Variety of Ways Different Services.
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Advisor: Professor.
Computer System Architectures Computer System Software
Introduction To Windows Azure Cloud
1 NETE4631 Managing the Cloud and Capacity Planning Lecture Notes #8.
SEDA: An Architecture for Well-Conditioned, Scalable Internet Services
November , 2009SERVICE COMPUTATION 2009 Analysis of Energy Efficiency in Clouds H. AbdelSalamK. Maly R. MukkamalaM. Zubair Department.
Performance Concepts Mark A. Magumba. Introduction Research done on 1058 correspondents in 2006 found that 75% OF them would not return to a website that.
Profile Driven Component Placement for Cluster-based Online Services Christopher Stewart (University of Rochester) Kai Shen (University of Rochester) Sandhya.
SAMANVITHA RAMAYANAM 18 TH FEBRUARY 2010 CPE 691 LAYERED APPLICATION.
Cloud Computing John Engates CTO, Rackspace Presented: Rackspace Customer Conference, 2008 October 29, 2008.
Web Search Using Mobile Cores Presented by: Luwa Matthews 0.
Frontiers in Massive Data Analysis Chapter 3.  Difficult to include data from multiple sources  Each organization develops a unique way of representing.
1 Challenges in Scaling E-Business Sites  Menascé and Almeida. All Rights Reserved. Daniel A. Menascé Department of Computer Science George Mason.
Distributed Information Systems. Motivation ● To understand the problems that Web services try to solve it is helpful to understand how distributed information.
1 Admission Control and Request Scheduling in E-Commerce Web Sites Sameh Elnikety, EPFL Erich Nahum, IBM Watson John Tracey, IBM Watson Willy Zwaenepoel,
Embedded System Lab. 정범종 A_DRM: Architecture-aware Distributed Resource Management of Virtualized Clusters H. Wang et al. VEE, 2015.
System Scalability. 1. General Observations The choice of platform for an application should consider the ability to grow the application with more users.
Ó 1998 Menascé & Almeida. All Rights Reserved.1 Part V Workload Characterization for the Web.
Measuring the Capacity of a Web Server USENIX Sympo. on Internet Tech. and Sys. ‘ Koo-Min Ahn.
June 30 - July 2, 2009AIMS 2009 Towards Energy Efficient Change Management in A Cloud Computing Environment: A Pro-Active Approach H. AbdelSalamK. Maly.
Capacity Planning Plans Capacity Planning Operational Laws
Revision - 01 Intel Confidential Page 1 Intel HPC Update Norfolk, VA April 2008.
Internet Applications: Performance Metrics and performance-related concepts E0397 – Lecture 2 10/8/2010.
MidVision Enables Clients to Rent IBM WebSphere for Development, Test, and Peak Production Workloads in the Cloud on Microsoft Azure MICROSOFT AZURE ISV.
CISC 849 : Applications in Fintech Namami Shukla Dept of Computer & Information Sciences University of Delaware iCARE : A Framework for Big Data Based.
Scalability == Capacity * Density.
1 PerfCenter and AutoPerf: Tools and Techniques for Modeling and Measurement of the Performance of Distributed Applications Varsha Apte Faculty Member,
Introduction to Performance Testing Performance testing is the process of determining the speed or effectiveness of a computer, network, software program.
Background Computer System Architectures Computer System Software.
REMINDER Check in on the COLLABORATE mobile app Oracle Performance Management with vCenter Operations Manager and Oracle Enterprise Manager (OEM) Adapter.
© 2008 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice ProLiant G5 to G6 Processor Positioning.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Designing a Grid Computing Architecture: A Case Study of Green Computing Implementation Using SAS® N.Krishnadas Indian Institute of Management, Kozhikode.
Chapter 8 Environments, Alternatives, and Decisions.
Lecture 2: Performance Evaluation
Recipes for Use With Thin Clients
Software Architecture in Practice
Green cloud computing 2 Cs 595 Lecture 15.
Software Architecture in Practice
Implementing a Load-balancing Web Server Using Red Hat Cluster Suite
File Transfer Issues with TCP Acceleration with FileCatalyst
Presentation & Demo August 7, 2018 Bill Shelden.
Admission Control and Request Scheduling in E-Commerce Web Sites
Software models - Software Architecture Design Patterns
Operating Systems : Overview
Operating Systems : Overview
SAMANVITHA RAMAYANAM 18TH FEBRUARY 2010 CPE 691
Computer Evolution and Performance
Software Acceleration in Hybrid Systems Xiaoqiao (XQ) Meng IBM T. J
Beyond FTP & hard drives: Accelerating LAN file transfers
Presentation transcript:

Alternative Performance Metrics for Server RFPs Joe Temple Low Country North Shore Consulting

Local Factors / Constraints Non-Functional Requirements Technology Adoption Strategic Direction Cost Models Reference Architectures System z System x Power Workload Fit This is an IBM Chart that bridges from platform selection into Performance Architecture

Fit for Purpose Workload Types Mixed Workload – Type 1 Scales up Updates to shared data and work queues Complex virtualization Business Intelligence with heavy data sharing and ad hoc queries Parallel Data Structures – Type 3 Small Discrete – Type 4 Application Function Data Structure Usage Pattern SLA Integration Scale Highly Threaded – Type 2 Scales well on clusters XML parsing Buisness intelligence with Structured Queries HPC applications Scales well on large SMP Web application servers Single instance of an ERP system Some partitioned databases Limited scaling needs HTTP servers File and print FTP servers Small end user apps Black are design factors Blue are local factors This is the IBM preSales Architects ‘ view of workload types

Fitness Parameters in Machine Design Can customized to machines of interest. Need to know the specific comparisons desired These parameters were chosen to represent the ability to handle, parallel, serial and bulk data traffic. This is based on Greg Pfister’s work on workload characterization in In Search of CLusters

Definitions TP - Thread Speed X Threads Thread Speed ~ Adjusted Clock Rate ITR - Internal Throughput Rate Peak rate as measured in benchmarks ITR <= TP ETR – External Throughput Rate Average rate as delivered in production ETR ~ ITR X Average Utilization

Throughput, Saturation, Capacity 6 TPMeasured ITRCapacity TP  Pure Parallel CPUITR  Other resources and Serialization ETR  Load and Response Time

Very, Very Few Clients experience ITR Most enterprises are interested in ETR ~ Average Utilization X ITR Most users experience response time

Throughput 8 Throughput: TP (Assume parallel load with no thread interactions) Saturation: Internal Throughput Rate (ITR) ITR  TP when highly parallel throughput is not limited by “other” resources (I/O, Memory, Bandwidth, Software, Cache) Capacity: External Throughput Rate (ETR) Utilization limited to meet response time.

Effect of using single dimension metrics. (Max Machines) 9 The “standard metrics” do not leverage cache. This leads to the pure ITR view of relative capacity on the right. Common Metrics: ITR  TP ETR  ITR Power advantaged z is not price competitive Consolidation: ETR << ITR unless loads are consolidated Consolidation accumulates working sets Power and z advantaged Cache can also mitigate “Saturation”

Typical x86 Consolidation 8X work on 4X CPUs  2X Average 39%, Peak 76% Peak to Average = 1.95 Average 61%, Peak 78% Peak to Average = 1.28 Enterprise Server Consolidation 64X work on 18X CPUs  3.6 X Average 21%, Peak 79% Peak to Average = 3.76 Dedicated x86 Server 1 X work on 1X CPUs  1 X Consolidation

The Math Behind Consolidation Roger’s Equation: Uavg = 1/(1+HR(avg)) Where HR(avg) = kcN 1/2 For Distribution of work: N = s (the number of servers per load) For Consolidation of work: N =1/ n (the number of loads per server) k is a design parameter (Service Level) c is the variability of the initial load

Response Time and Variability 12 Acceptable Response Time Hi Variability Moderate Variability Low Variability “No Variability”

The math behind the Hockey Stick Use your favorite queuing model. If you use M/M/1 or M/M/K models cSQRT(N) will be assumed to be 1. We used an estimator for M/G/1 or G/G/1 T = To(1+ c 2 N(u/(1-u)) Notice that elements of Rogers’ equation appear In both cases N affects the variability impact We also know that HR(u) = (1-u)/u T = To(1+ c 2 N/HR(u))

We have a model which uses these concepts. It generates characteristic curves And profiles machinesAnd profiles machines

Bottom Line on workload fit “Best” is user dependent – Some dependence on “workload factors” – Mostly dependent on parallelism, size, usage pattern and service level of loads – Small, variable loads will lean toward density – Larger, more steady loads will lean toward throughput – Need to decide figure(s) of merit Designers should set at least 2 requirements: – Throughput and Thread Capacity – ETR and Density – Density and Response Time – Etc.

Comparing Max Machines One core per socket of Power7 is dedicated to VIO and Intel pathlength is penalized for I/O

What is the figure of merit? ITR – What we benchmark? ETR – closer to business value ($/Day)? Average Response Time – User experience? Response time at Peak – speed at max load? Stack Density – VMs/Core (Loads per core)? Average Utilization – Efficiency of use? None of the machines is “best across the board” Designers should specify at least 2 metrics

Stacked single thread workloads Max Threads Consolidation Model Response Time Parallelism Serial1 Threads1 SLA and Distributed Variability k3.1 c2 Ndist1 Each workload small and variable. Z has highest density and highest speed Power has highest throughput (SMT4)

Bigger, more Parallel Loads Max Threads Consolidation Model Response Time Parallelism Serial0.1 Threads16 SLA and Distributed Variability k3.1 c1 Ndist1 Moderate Variability, Larger workloads Power still has highest throughput z has less speed advantage z maintains density advantage

Very Large Parallel Loads Max Threads Consolidation Model Response Time Parallelism Serial0.01 Threads64 SLA and Distributed Variability k3.1 c.25 Ndist1 Low Variability, Larger workloads Power is clear winner except for density

Low Country North Shore Consulting Visit lc-ns.com or Joe at

lc-ns work research and services Collateral Development and tech writing Further development of workload fit model Application of workload fit model to specific comparisons (will not compete with IBM). Specification and application of benchmarks to model Understanding tails of short interval utilization distributions Validation of sizings Machine positioning Workload analysis (usage patterns, response time parallelism and load consolidation/distribution.) Skill transfer/ Education / Speaking on the above Analysis/Development of Intellectual Property Leadership Mentoring / Coaching