ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 31 Alternative System Description If all w k are given initially as Then,

Slides:



Advertisements
Similar presentations
Many useful applications, especially in queueing systems, inventory management, and reliability analysis. A connection between discrete time Markov chains.
Advertisements

Linear Programming (LP) (Chap.29)
Chapter 4 Retiming.
Dynamic Decision Processes
In this handout Stochastic Dynamic Programming
TCOM 501: Networking Theory & Fundamentals
Lecture 13 – Continuous-Time Markov Chains
048866: Packet Switch Architectures Dr. Isaac Keslassy Electrical Engineering, Technion Review.
Math443/543 Mathematical Modeling and Optimization
EMGT 501 Fall 2005 Midterm Exam SOLUTIONS.
TCOM 501: Networking Theory & Fundamentals
Background Material: Markov Decision Process. Reference Class notes Further studies: Dynamic programming and Optimal Control D. Bertsekas, Volume 1 Chapters.
Markov Chains Chapter 16.
Chapter 18 Management of Waiting Lines
Chapter 10 Dynamic Programming. 2 Agenda for This Week Dynamic Programming –Definition –Recursive Nature of Computations in DP –Forward and Backward Recursion.
Max-flow/min-cut theorem Theorem: For each network with one source and one sink, the maximum flow from the source to the destination is equal to the minimal.
Introduction to Queuing Theory
Flows and Networks Plan for today (lecture 5): Last time / Questions? Waiting time simple queue Little Sojourn time tandem network Jackson network: mean.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 1 Chapter 1: The DP Algorithm To do:  sequential decision-making  state.
Flows and Networks Plan for today (lecture 5): Last time / Questions? Blocking of transitions Kelly / Whittle network Optimal design of a Kelly / Whittle.
1 Chapter 5 Flow Lines Types Issues in Design and Operation Models of Asynchronous Lines –Infinite or Finite Buffers Models of Synchronous (Indexing) Lines.
An Overview of Dynamic Programming Seminar Series Joe Hartman ISE October 14, 2004.
Markov Decision Processes1 Definitions; Stationary policies; Value improvement algorithm, Policy improvement algorithm, and linear programming for discounted.
Decision Making in Robots and Autonomous Agents Decision Making in Robots and Autonomous Agents The Markov Decision Process (MDP) model Subramanian Ramamoorthy.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 49 DP can give complete quantitative solution Example 1: Discrete, finite.
1 1 Slide © 2000 South-Western College Publishing/ITP Slides Prepared by JOHN LOUCKS.
Networks of Queues Plan for today (lecture 6): Last time / Questions? Product form preserving blocking Interpretation traffic equations Kelly / Whittle.
1 Queueing Theory Frank Y. S. Lin Information Management Dept. National Taiwan University
TexPoint fonts used in EMF.
1 Networks of queues Networks of queues reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity,
Dynamic Programming Applications Lecture 6 Infinite Horizon.
Network Design and Analysis-----Wang Wenjie Queuing Theory III: 1 © Graduate University, Chinese academy of Sciences. Network Design and Performance Analysis.
CDA6530: Performance Models of Computers and Networks Chapter 7: Basic Queuing Networks TexPoint fonts used in EMF. Read the TexPoint manual before you.
1 1 © 2003 Thomson  /South-Western Slide Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
Markov Chains X(t) is a Markov Process if, for arbitrary times t1 < t2 < < tk < tk+1 If X(t) is discrete-valued If X(t) is continuous-valued i.e.
Chapter 61 Continuous Time Markov Chains Birth and Death Processes,Transition Probability Function, Kolmogorov Equations, Limiting Probabilities, Uniformization.
Flows and Networks Plan for today (lecture 6): Last time / Questions? Kelly / Whittle network Optimal design of a Kelly / Whittle network: optimisation.
Chapter 1 Introduction. “Wait-in-line” is a common phenomenon in everywhere. Reason: Demand is more than service. “How long must a customer wait?” or.
1 Inventory Control with Time-Varying Demand. 2  Week 1Introduction to Production Planning and Inventory Control  Week 2Inventory Control – Deterministic.
IT Applications for Decision Making. Operations Research Initiated in England during the world war II Make scientifically based decisions regarding the.
Model 5 Long Distance Phone Calls By Benjamin Cutting
Thursday, May 2 Dynamic Programming – Review – More examples Handouts: Lecture Notes.
Chapter 6 Product-Form Queuing Network Models Prof. Ali Movaghar.
Queuing Theory.  Queuing Theory deals with systems of the following type:  Typically we are interested in how much queuing occurs or in the delays at.
1 1 © 2003 Thomson  /South-Western Slide Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
DEPARTMENT/SEMESTER ME VII Sem COURSE NAME Operation Research Manav Rachna College of Engg.
11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.
Review for E&CE Find the minimal cost spanning tree for the graph below (where Values on edges represent the costs). 3 Ans. 18.
Chapter 5 Elementary Stochastic Analysis Prof. Ali Movaghar.
Flows and Networks Plan for today (lecture 6): Last time / Questions? Kelly / Whittle network Optimal design of a Kelly / Whittle network: optimisation.
Chapter 1 Introduction.
Water Resources Planning and Management Daene McKinney
Analytics and OR DP- summary.
Simulation Examples STOCHASTIC PROCESSES.
6.5 Stochastic Prog. and Benders’ decomposition
Lecture on Markov Chain
Finite M/M/1 queue Consider an M/M/1 queue with finite waiting room.
Dynamic Programming Lecture 13 (5/31/2017).
Chapter 6. Large Scale Optimization
Dynamic Programming General Idea
Flows and Networks Plan for today (lecture 6):
Not Always Feasible 2 (3,5) 1 s (0,2) 3 (, u) t.
Introduction to Queueing Theory
Dynamic Programming General Idea
Dynamic Programming II DP over Intervals
6.5 Stochastic Prog. and Benders’ decomposition
Discrete-time markov chain (continuation)
CS723 - Probability and Stochastic Processes
Chapter 6. Large Scale Optimization
Presentation transcript:

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 31 Alternative System Description If all w k are given initially as Then, a system equation can be obtained as  System transition law  system equation

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 32 Alternative System Description System equation: better for models with continuous state and control spaces (e.g. S k =  ). Transition law: better for models with discrete state and control spaces (e.g. S k = {0,1,2, …, k}).

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 33 Alternative System Description Example: Integer, Finite Capacity Inventory Control Problem x k, u k : non negative integers. ( x k + u k )  2 : Maximum capacity is 2 units. S k = {0, 1, 2} C k = {0, 1, 2} Excess demand max{0, (w k -x k -u k )} is lost. D k ={0, 1, …}

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 34 Alternative System Description System (stock) equation to keep  2, put constraints on u k

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 35 Alternative System Description Assume Prob{w k =0}=0.1 Prob{w k =1}=0.7 Prob{w k =2}=0.2 Let’s find the (time independent) state transition law

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 36 Alternative System Description u k = 0 p 00 = 1, p 01 = p 02 = 0 p 10 = Prob{w k  1} = = 0.9 p 11 = Prob{w k = 0} = 0.1; p 12 = 0 p 20 = Prob{w k = 2} = 0.2 p 21 = Prob{w k = 1} = 0.7 p 22 = Prob{w k = 0} = 0.1

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 37 Alternative System Description stock = 2 stock = 1 stock = 0 stock = 2 stock = 1 stock = Transition Probability Diagram ( u k =0 ) k k+1

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 38 Alternative System Description stock = 2 stock = 1 stock = 0 stock = 2 stock = 1 stock = Transition Probability Diagram ( u k =1 ) k k Could put arbitrary numbers, then constraint u k * * : p 11 = Prob{w k = 0} = 0.7

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 39 Alternative System Description stock = 2 stock = 1 stock = 0 stock = 2 stock = 1 stock = Transition Probability Diagram ( u k =2 ) k k

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 40 Alternative System Description Example: Finite Capacity Queue with variable service rate. arrival departure server 0 k-1k k+1 N nn w k-1 w k+1 = arrivals Service can start/end only at the beginning (end) of the period. Customers finding system full depart and don’t come back. x k : queue length, 0 ≤ x k ≤ n (finite capacity) u k =

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 41 w k : # customer that arrive during period k if queue is full, customers leave and don’t return; {w k } i.i.d. p m : probability of m arrivals. q f > q s : probability of service complete in one period q f = 0.8, q s = 0.3 P( x k+1 = j │ x k = i, u k = F ) := p ij Alternative System Description

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 42 Alternative System Description u k = F i = 0 p 0j = p j, 0 ≤ j ≤ n - 1 i > 0 j < i -1 p ij = 0 j = i -1 P i i - 1 = Prob {service completed, no arrivals} = q f p 0 at most 1 customer serve in one period

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 43 u k = F i – 1 < j < n -1 p ij = Prob {service completed, j – i + 1 arrivals } + Prob {service not completed, j – i arrivals } = q f p j – i ( 1 - q f ) p j - i j = n -1 j = n Alternative System Description

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 44 The Dynamic Programming Algorithm Example: (Deterministic) Shortest Path stages Start End

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 45 The Dynamic Programming Algorithm (i)Forward, myopic minimization: start at node 0, and proceed forward, minimization the length (cost) for present stage:  0  1 1  3 1  6 6   Not optimal, since  0  2 2  5 5  7 7 

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 46 The Dynamic Programming Algorithm What we need to do is to minimize costs for current stage and the path these after. (x k, u k )  x k+1 Shortest path from x k+1 to end (ii) Backward minimization: stage 3 6  8 (6) 7  8 (2)

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 47 stage 2 6  8  6 (13) The Dynamic Programming Algorithm 4  7  8 (5) 5  7  8 (6) stage 1 6  8  7  8 (10) 2  4  7  8 (6) stage 0 : 2  4  7  8 (10)  J*(0) = 10  * = { 2, 4, 7, 8 }

ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 48 The Dynamic Programming Algorithm Subproblem: Suppose we start now at node 2, what is the shortest path? A/  * = { 4, 7, 8 } :truncated optimal solution for original problem