On the convergence of SDDP and related algorithms Speaker: Ziming Guan Supervisor: A. B. Philpott Sponsor: Fonterra New Zealand.

Slides:



Advertisements
Similar presentations
ONS SDDP Workshop, August 17, 2011 Slide 1 of 31 Andy Philpott EPOC ( joint work with Ziming Guan (now at UBC/BC Hydro) Electricity Market.
Advertisements

Wind power scenario generation Geoffrey Pritchard University of Auckland by regression clustering.
MINIMUM COST FLOWS: NETWORK SIMPLEX ALGORITHM A talk by: Lior Teller 1.
 1  Outline  terminating and non-terminating systems  theories for output analysis  Strong Law of Large Numbers  Central Limit Theorem  Regenerative.
ICCV 2007 tutorial Part III Message-passing algorithms for energy minimization Vladimir Kolmogorov University College London.
Modelling Developments at Power Systems Research Tom Halliburton EPOC Meeting 9 th July 2014.
UNINFORMED SEARCH Problem - solving agents Example : Romania  On holiday in Romania ; currently in Arad.  Flight leaves tomorrow from Bucharest.
CHAPTER 2 D IRECT M ETHODS FOR S TOCHASTIC S EARCH Organization of chapter in ISSO –Introductory material –Random search methods Attributes of random search.
Separating Hyperplanes
EPOC Optimization Workshop, July 8, 2011 Slide 1 of 41 Andy Philpott EPOC ( joint work with Anes Dallagi, Emmanuel Gallet, Ziming Guan.
Applications of Stochastic Programming in the Energy Industry Chonawee Supatgiat Research Group Enron Corp. INFORMS Houston Chapter Meeting August 2, 2001.
ONS SDDP Workshop, August 17, 2011 Slide 1 of 50 Andy Philpott Electric Power Optimization Centre (EPOC) University of Auckland (
Planning under Uncertainty
1 Logic-Based Methods for Global Optimization J. N. Hooker Carnegie Mellon University, USA November 2003.
Simulation Where real stuff starts. ToC 1.What, transience, stationarity 2.How, discrete event, recurrence 3.Accuracy of output 4.Monte Carlo 5.Random.
MAE 552 – Heuristic Optimization Lecture 6 February 6, 2002.
Job Release-Time Design in Stochastic Manufacturing Systems Using Perturbation Analysis By: Dongping Song Supervisors: Dr. C.Hicks & Dr. C.F.Earl Department.
Theoretical Results on Base Station Movement Problem for Sensor Network Yi Shi ( 石毅 ) and Y. Thomas Hou ( 侯一釗 ) Virginia Tech, Dept. of ECE IEEE Infocom.
A Finite Sample Upper Bound on the Generalization Error for Q-Learning S.A. Murphy Univ. of Michigan CALD: February, 2005.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Optimal Fixed-Size Controllers for Decentralized POMDPs Christopher Amato Daniel.
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
Reinforcement Learning: Learning algorithms Yishay Mansour Tel-Aviv University.
Modelling inflows for SDDP Dr. Geoffrey Pritchard University of Auckland / EPOC.
Experts and Boosting Algorithms. Experts: Motivation Given a set of experts –No prior information –No consistent behavior –Goal: Predict as the best expert.
Hydro Optimization Tom Halliburton. Variety Stochastic Deterministic Linear, Non-linear, dynamic programming Every system is different Wide variety.
© Copyright McGraw-Hill CHAPTER 6 The Normal Distribution.
Polyhedral Risk Measures Vadym Omelchenko, Institute of Information Theory and Automation, Academy of Sciences of the Czech Republic.
Computational Stochastic Optimization: Bridging communities October 25, 2012 Warren Powell CASTLE Laboratory Princeton University
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 1 Chapter 1: The DP Algorithm To do:  sequential decision-making  state.
Adaptive CSMA under the SINR Model: Fast convergence using the Bethe Approximation Krishna Jagannathan IIT Madras (Joint work with) Peruru Subrahmanya.
Linear Programming Topics General optimization model LP model and assumptions Manufacturing example Characteristics of solutions Sensitivity analysis Excel.
EPOC Winter Workshop, September 5, 2008 Andy Philpott The University of Auckland (joint work with Kailin Lee, Golbon Zakeri)
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
Quasi-static Channel Assignment Algorithms for Wireless Communications Networks Frank Yeong-Sung Lin Department of Information Management National Taiwan.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 49 DP can give complete quantitative solution Example 1: Discrete, finite.
Electricity markets, perfect competition and energy shortage risks Andy Philpott Electric Power Optimization Centre University of.
1 Optimization Based Power Generation Scheduling Xiaohong Guan Tsinghua / Xian Jiaotong University.
Simulated Annealing.
EPOC Winter Workshop, October 26, 2010 Slide 1 of 31 Andy Philpott EPOC ( joint work with Vitor de Matos, Ziming Guan Advances in DOASA.
Algorithms for MAP estimation in Markov Random Fields Vladimir Kolmogorov University College London.
Exact and heuristics algorithms
Presented By Dr. Mohsen Alardhi College of Technological Studies, Kuwait April 19 th,2009.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
Algorithms 1.Notion of an algorithm 2.Properties of an algorithm 3.The GCD algorithm 4.Correctness of the GCD algorithm 5.Termination of the GCD algorithm.
Linear Program Set Cover. Given a universe U of n elements, a collection of subsets of U, S = {S 1,…, S k }, and a cost function c: S → Q +. Find a minimum.
1 Monte-Carlo Planning: Policy Improvement Alan Fern.
Stochastic processes for hydrological optimization problems
Stochastic Optimization
Resource Allocation in Hospital Networks Based on Green Cognitive Radios 王冉茵
Stochastic Optimization for Markov Modulated Networks with Application to Delay Constrained Wireless Scheduling Michael J. Neely University of Southern.
September 28, 2000 Improved Simultaneous Data Reconciliation, Bias Detection and Identification Using Mixed Integer Optimization Methods Presented by:
5.3 Mixed Integer Nonlinear Programming Models. A Typical MINLP Model.
Application of Dynamic Programming to Optimal Learning Problems Peter Frazier Warren Powell Savas Dayanik Department of Operations Research and Financial.
Linear Programming Piyush Kumar Welcome to CIS5930.
Approximation Algorithms based on linear programming.
1 Maximum Flows CONTENTS Introduction to Maximum Flows (Section 6.1) Introduction to Minimum Cuts (Section 6.1) Applications of Maximum Flows (Section.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Achieving Goals in Decentralized POMDPs Christopher Amato Shlomo Zilberstein UMass.
5.3 Mixed-Integer Nonlinear Programming (MINLP) Models
Analytics and OR DP- summary.
Morgan Bruns1, Chris Paredis1, and Scott Ferson2
5.3 Mixed Integer Nonlinear Programming Models
Chapter 6. Large Scale Optimization
Stochastic Network Interdiction
Department of Information Management National Taiwan University
CS 188: Artificial Intelligence Fall 2007
Hidden Markov Models (cont.) Markov Decision Processes
Chapter 6. Large Scale Optimization
Presentation transcript:

On the convergence of SDDP and related algorithms Speaker: Ziming Guan Supervisor: A. B. Philpott Sponsor: Fonterra New Zealand

Motivation Pereira and Pinto, Multi-Stage Stochastic Optimization Applied to Energy Planning, Mathematical Programming, 52, pp , 1991.

Summary Description of problem class SDDP and its related algorithm Theoretical convergence Implementation issues

Properties for random quantities Random quantities appear only on the right-hand side of the linear constraints in each stage. The set of random outcomes is discrete and finite. Random quantities in different stages are independent. Can accommodate PARMA process for RHS uncertainty.

                           p 12 p 11 p 13 p 21 Scenario tree, scenario outcome, scenario

Hydro-thermal scheduling

Stage problem

Cuts Θ (t+1) Reservoir storage, x(t+1)

Stochastic Dual Dynamic Programming [Pereira and Pinto, 1991] Initialization: Sample some scenarios and fix them through the course of the algorithm. Forward pass: For stage t=1,…,T, solve the stage t problem for each scenario. Calculate the lower bound and upper bound. If not converge, –Backward pass: For stage t=T-1,…,1, for the stage t problem in each scenario, solve all stage t+1 problems to calculate a cut for stage t problems. –Back to Forward pass.

                  p 11 p 13          p 12

         p 11 p 13 p 12

                  p 11 p 13          p 12

Dynamic Outer Approximation Sampling Algorithm No upper bound calculation until algorithm is terminated.

                  p 11 p 13        p 12

         p 11 p 13 p 12

                  p 11 p 13        p 12

We have a convergence proof for DOASA. This can be used to understand the convergence behaviour of SDDP.

Sampling properties of DOASA Forward Pass Sampling Property (FPSP): Each scenario is traversed infinitely many times with probability 1 in the forward pass.

How do we guarantee this? Either Independently sample a single outcome in each stage with a positive probability for each scenario outcome in the forward pass. Repeat an exhaustive enumeration of each scenario in the forward pass.

Convergence Theorem Under FPSP, DOASA converges with probability 1 to an optimal solution to the stage 1 problem in a finite number of iterations.

Sampling in cut calculation Sample some stage problems. Keep a list of dual solutions, search the best one for the stage problem that are not sampled. Backward Pass Sampling Property (BPSP): In any stage, each scenario outcome is visited infinitely many times with probability 1 in the backward pass.

Convergence Theorem Under FPSP and BPSP, the algorithm converges with probability 1 to an optimal solution to the stage 1 problem in a finite number of iterations.

Corollaries If every outcome is used in cut calculation we only need FPSP. We can bias sampling as long as FPSP is satisfied. (Note estimation of upper bound needs unbiased scenarios.)

Resampling SDDP does not resample the forward pass. It creates N scenarios of inflows at the start. FPSP is NOT satisfied. SDDP will terminate with probability 1. Cuts give a lower bound, but policy need not be optimal.

Dry Wet   Always Dry, when at convergence...

Negative inflows SDDP uses PARMA model for inflows. Negative inflows might result – not physically possible. Some implementations adjust random outcomes to make inflow non-negative – this destroys stage-wise independence. Cut sharing is no longer valid. Log-normal inflows not valid for convexity reasons.

Convexity matters in backward pass Transmission losses can make stage problem not convex if free disposal is not allowed. Unit commitment integer effects are not convex.

Convergence expectation We run DOASA on a problem at Fonterra NZ. Maximum size for convergence = 12 stages x 24 states. In revenue management application, 8 states, 5000 stages converge, 20 states, 5000 stages does not. Convergence is problem dependent.

Case study: NZ model S N demand TPOHAWMAN

Computational results: NZ model 9 reservoirs 52 weekly stages 30 inflow outcomes per stage Model written in AMPL/CPLEX Takes 100 iterations and 2 hours on a standard Windows PC to converge

policy simulated with historical inflow sequences

END