An Optimal Service Ordering for a World Wide Web Server A Presentation for the Fifth INFORMS Telecommunications Conference March 6, 2000 Amy Csizmar Dalal.

Slides:



Advertisements
Similar presentations
Thrasyvoulos Spyropoulos / Eurecom, Sophia-Antipolis 1  Load-balancing problem  migrate (large) jobs from a busy server to an underloaded.
Advertisements

Lab Assignment 1 COP 4600: Operating Systems Principles Dr. Sumi Helal Professor Computer & Information Science & Engineering Department University of.
Short-Term Fairness and Long- Term QoS Lei Ying ECE dept, Iowa State University, Joint work with Bo Tan, UIUC and R. Srikant, UIUC.
Load Balancing of Elastic Traffic in Heterogeneous Wireless Networks Abdulfetah Khalid, Samuli Aalto and Pasi Lassila
Continuous Time Markov Chains and Basic Queueing Theory
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
Looking at the Server-side of P2P Systems Yi Qiao, Dong Lu, Fabian E. Bustamante and Peter A. Dinda Department of Computer Science Northwestern University.
Cs238 CPU Scheduling Dr. Alan R. Davis. CPU Scheduling The objective of multiprogramming is to have some process running at all times, to maximize CPU.
Performance Evaluation
MGTSC 352 Lecture 23: Congestion Management Introduction: Asgard Bank example Simulating a queue Types of congested systems, queueing template Ride’n’Collide.
Little’s Theorem Examples Courtesy of: Dr. Abdul Waheed (previous instructor at COE)
OS Fall ’ 02 Performance Evaluation Operating Systems Fall 2002.
Location Models For Airline Hubs Behaving as M/D/C Queues By: Shuxing Cheng Yi-Chieh Han Emile White.
Dimitrios Konstantas, Evangelos Grigoroudis, Vassilis S. Kouikoglou and Stratos Ioannidis Department of Production Engineering and Management Technical.
Decentralised load balancing in closed and open systems A. J. Ganesh University of Bristol Joint work with S. Lilienthal, D. Manjunath, A. Proutiere and.
1 Chapter 5 Flow Lines Types Issues in Design and Operation Models of Asynchronous Lines –Infinite or Finite Buffers Models of Synchronous (Indexing) Lines.
MIT Fun queues for MIT The importance of queues When do queues appear? –Systems in which some serving entities provide some service in a shared.
Lecture 10: Queueing Theory. Queueing Analysis Jobs serviced by the system resources Jobs wait in a queue to use a busy server queueserver.
NETE4631:Capacity Planning (2)- Lecture 10 Suronapee Phoomvuthisarn, Ph.D. /
1 Call Admission Control for Multimedia Services in Mobile Cellular Networks : A Markov Decision Approach Jihyuk Choi; Taekyoung Kwon; Yanghee Choi; Naghshineh,
Manijeh Keshtgary. Queuing Network: model in which jobs departing from one queue arrive at another queue (or possibly the same queue)  Open and Closed.
Queuing Theory Basic properties, Markovian models, Networks of queues, General service time distributions, Finite source models, Multiserver queues Chapter.
Modeling and Analysis of Computer Networks
Queuing Networks Jean-Yves Le Boudec 1. Networks of Queues Stability Queuing networks are frequently used models The stability issue may, in general,
1 Chapters 8 Overview of Queuing Analysis. Chapter 8 Overview of Queuing Analysis 2 Projected vs. Actual Response Time.
Scheduling Periodic Real-Time Tasks with Heterogeneous Reward Requirements I-Hong Hou and P.R. Kumar 1 Presenter: Qixin Wang.
SIMULATION OF A SINGLE-SERVER QUEUEING SYSTEM
Carnegie Mellon University Computer Science Department 1 OPEN VERSUS CLOSED: A CAUTIONARY TALE Bianca Schroeder Adam Wierman Mor Harchol-Balter Computer.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems with Multi-programming Chapter 4.
“A cost-based admission control algorithm for digital library multimedia systems storing heterogeneous objects” – I.R. Chen & N. Verma – The Computer Journal.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
CS 450 – Modeling and Simulation Dr. X. Topics System Analysis Arrival process The Queue The Server The simulation.
Flows and Networks Plan for today (lecture 3): Last time / Questions? Output simple queue Tandem network Jackson network: definition Jackson network: equilibrium.
Mohammad Khalily Islamic Azad University.  Usually buffer size is finite  Interarrival time and service times are independent  State of the system.
Introduction to Operating System Created by : Zahid Javed CPU Scheduling Fifth Lecture.
Lecture 4 CPU scheduling. Basic Concepts Single Process  one process at a time Maximum CPU utilization obtained with multiprogramming CPU idle :waiting.
CPU scheduling.  Single Process  one process at a time  Maximum CPU utilization obtained with multiprogramming  CPU idle :waiting time is wasted 2.
Dynamic Resource Allocation for Shared Data Centers Using Online Measurements By- Abhishek Chandra, Weibo Gong and Prashant Shenoy.
Networks (3TU) Plan for today (lecture 5): Last time / Questions? Tandem network Jackson network: definition Jackson network: equilibrium distribution.
Renewal Theory Definitions, Limit Theorems, Renewal Reward Processes, Alternating Renewal Processes, Age and Excess Life Distributions, Inspection Paradox.
OPERATING SYSTEMS CS 3502 Fall 2017
OPERATING SYSTEMS CS 3502 Fall 2017
CPU Scheduling CSSE 332 Operating Systems
Scheduling and Fairness
The Impact of Replacement Granularity on Video Caching
Dynamic Graph Partitioning Algorithm
Deadline Scheduling and Heavy tail distributionS
Scheduling (Priority Based)
Chapter 6: CPU Scheduling
CPU Scheduling G.Anuradha
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
Queuing models Basic definitions, assumptions, and identities
Queuing models Basic definitions, assumptions, and identities
3: CPU Scheduling Basic Concepts Scheduling Criteria
Chapter5: CPU Scheduling
Chapter 6: CPU Scheduling
CPU SCHEDULING.
Chapter 5: CPU Scheduling
Chapter 5: CPU Scheduling
Javad Ghaderi, Tianxiong Ji and R. Srikant
Uniprocessor scheduling
Operating System , Fall 2000 EA101 W 9:00-10:00 F 9:00-11:00
Chapter 6: CPU Scheduling
Module 5: CPU Scheduling
Chapter 6: CPU Scheduling
Resource Sharing with Subexponential Distributions
Don Porter Portions courtesy Emmett Witchel
Module 5: CPU Scheduling
VIRTUE MARYLEE MUGURACHANI QUEING THEORY BIRTH and DEATH.
Presentation transcript:

An Optimal Service Ordering for a World Wide Web Server A Presentation for the Fifth INFORMS Telecommunications Conference March 6, 2000 Amy Csizmar Dalal Northwestern University Department of Electrical and Computer Engineering Scott Jordan University of California, Irvine Department of Electrical and Computer Engineering

Overview ·Motivation/problem statement ·Derivation of optimal policy ·Two extensions to the original model ·Simulation results ·Conclusions

Problems with current web server operation ·Currently, web servers share processor resources among several requests simultaneously. ·Web users are impatient--they tend to abort page requests that do not get a response within several seconds. ·If users time out, the server wastes resources on requests that never complete-- a problem especially for heavily-loaded web servers. ·Objective: decrease user-perceived latency by minimizing the number of user file requests that “time out” before completion

A queueing model approach to measuring web server performance ·Performance = P{user will remain at the server until its service completes} ·Decreasing function of response time ·Service times ~ exp(  ), i.i.d, proportional to sizes of requested files ·Request arrivals ~ Poisson( ) ·Server is unaware if/when requests are aborted ·Server alternates between busy cycles, when there is at least one request in the system, and idle cycles (state 0), when there are no requests waiting or in service. These cycles are i.i.d. ·Switching time between requests is negligible

Web server performance model Revenue earned by the server under policy p from serving request i to completion: By the Renewal Reward Theorem, the average reward earned by the server per unit time under policy p is Potential revenue: g ip (t)  c ip

Web server performance: Objective Find a service policy that maximizes revenue earned per unit time:  optimal policy

Requirements for an optimal service ordering policy Lemma 1: An optimal policy is non-idling Difference in expected revenues = g j* (a 2 )  d (1-e -c d )

Requirements for an optimal service ordering policy Lemma 2: An optimal service ordering policy switches between jobs in service only upon an arrival to or departure from the system If g (j+1)* (a 1 )  g j* (a 1 ): use first alternate policy Revenue difference = g (j+1)* (a 1 )  d (1-e -c d ) If g (j+1)* (a 1 ) < g j* (a 1 ): use second alternate policy Revenue difference =  d e -c d (e -ca 1 - e -ca 2 ) (e cx j* - e cx (j+1)* )

Requirements for an optimal service ordering policy Lemma 3: An optimal policy is non-processor-sharing Proof: Consider a portion of the sample path over which the server processes two jobs to completion Three possibilities: ·serve 1 then 2: expected revenue is ·serve 2 then 1: expected revenue is ·serve both until one completes, then serve the other: expected revenue is

Requirements for an optimal service ordering policy Lemma 4: An optimal policy is Markov (independent of both past and future arrivals) Follows because service times and interarrival times are exponential

Definition of an optimal policy State vector at time t under policy p: best job: *The optimal policy chooses the best job to serve at any time t regardless of the order of service chosen for the rest of the jobs in the system (greedy)  LIFO-PR

Extension 1: Initial reward ·Initial reward C i : indicates user’s perceived probability of completion/service upon arrival at the web server ·Application: QoS levels ·Potential revenue: ·Natural extension of the original system model Optimal policy: serve the request that has the highest potential revenue (function of the time in the system so far and the initial reward)

Extension 2: Varying service time distributions ·Service time of request i ~ exp(  i ), independent ·Application: different content types best described by different distributions ·Additional initial assumptions: ›only switch between jobs in service upon an arrival to or departure from the system ›non-processor-sharing ·We show that the optimal policy is ›non-idling ›Markov

Extension 2 (continued) State vector at time t under policy p: Best job satisfies Optimal policy: Serve the request with the highest c i  i product

Simulation results ·Revenue (normalized by the number of requests served during a simulation run) as a function of offered load ( /  ) ·Mean file size (  ) = 2 kB ·Extension 1: C i ~ U[0,1) ·Extension 2:  1  2 kB,  2 = 14 kB,  3 = 100 kB ·Implemented in BONeS ™

Simulation results

Conclusions ·Using a greedy service policy rather than a fair policy at a web server can improve server performance by up to several orders of magnitude ·Performance improvements are most pronounced at high server loads ·Under certain conditions, a processor-sharing policy may generate a higher average revenue than comparable non- processor-sharing policies. These conditions remain to be quantified.