Linear & Nonlinear Programming -- Basic Properties of Solutions and Algorithms.

Slides:



Advertisements
Similar presentations
5.1 Real Vector Spaces.
Advertisements

Boyce/DiPrima 9th ed, Ch 2.8: The Existence and Uniqueness Theorem Elementary Differential Equations and Boundary Value Problems, 9th edition, by William.
Differential Equations Brannan Copyright © 2010 by John Wiley & Sons, Inc. All rights reserved. Chapter 08: Series Solutions of Second Order Linear Equations.
Introducción a la Optimización de procesos químicos. Curso 2005/2006 BASIC CONCEPTS IN OPTIMIZATION: PART II: Continuous & Unconstrained Important concepts.
Engineering Optimization
1 OR II GSLM Outline  classical optimization – unconstrained optimization  dimensions of optimization  feasible direction.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
Basic Feasible Solutions: Recap MS&E 211. WILL FOLLOW A CELEBRATED INTELLECTUAL TEACHING TRADITION.
Chapter 4: Network Layer
Visual Recognition Tutorial
Finance 30210: Managerial Economics Optimization.
MIT and James Orlin © Nonlinear Programming Theory.
THE MATHEMATICS OF OPTIMIZATION
Maximization without Calculus Not all economic maximization problems can be solved using calculus –If a manager does not know the profit function, but.
Constrained Maximization
15 PARTIAL DERIVATIVES.
INFINITE SEQUENCES AND SERIES
Finance 510: Microeconomic Analysis
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
Unconstrained Optimization Problem
Optimization using Calculus
Optimality Conditions for Nonlinear Optimization Ashish Goel Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
D Nagesh Kumar, IIScOptimization Methods: M2L3 1 Optimization using Calculus Optimization of Functions of Multiple Variables: Unconstrained Optimization.
Math for CSLecture 51 Function Optimization. Math for CSLecture 52 There are three main reasons why most problems in robotics, vision, and arguably every.
Computer Algorithms Mathematical Programming ECE 665 Professor Maciej Ciesielski By DFG.
THE MATHEMATICS OF OPTIMIZATION
Tier I: Mathematical Methods of Optimization
Managerial Economics Managerial Economics = economic theory + mathematical eco + statistical analysis.
Ch 8.1 Numerical Methods: The Euler or Tangent Line Method
1. The Simplex Method.
Linear Algebra Chapter 4 Vector Spaces.
SECOND-ORDER DIFFERENTIAL EQUATIONS
ME 2304: 3D Geometry & Vector Calculus Dr. Faraz Junejo Double Integrals.
1. Problem Formulation. General Structure Objective Function: The objective function is usually formulated on the basis of economic criterion, e.g. profit,
The mean value theorem and curve sketching
The importance of sequences and infinite series in calculus stems from Newton’s idea of representing functions as sums of infinite series.  For instance,
Simplex method (algebraic interpretation)
Nonlinear programming Unconstrained optimization techniques.
Pareto Linear Programming The Problem: P-opt Cx s.t Ax ≤ b x ≥ 0 where C is a kxn matrix so that Cx = (c (1) x, c (2) x,..., c (k) x) where c.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Nonlinear Programming Models
Section 2.3 Properties of Solution Sets
EASTERN MEDITERRANEAN UNIVERSITY Department of Industrial Engineering Non linear Optimization Spring Instructor: Prof.Dr.Sahand Daneshvar Submited.
Managerial Economics Managerial Economics = economic theory + mathematical eco + statistical analysis.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
Nonlinear Programming I Li Xiaolei. Introductory concepts A general nonlinear programming problem (NLP) can be expressed as follows: objective function.
12 INFINITE SEQUENCES AND SERIES. In general, it is difficult to find the exact sum of a series.  We were able to accomplish this for geometric series.
Chapter 4 Sensitivity Analysis, Duality and Interior Point Methods.
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Introduction to Optimization
Network Systems Lab. Korea Advanced Institute of Science and Technology No.1 Ch. 3 Iterative Method for Nonlinear problems EE692 Parallel and Distribution.
Boyce/DiPrima 9 th ed, Ch 11.3: Non- Homogeneous Boundary Value Problems Elementary Differential Equations and Boundary Value Problems, 9 th edition, by.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
Optimization in Engineering Design Georgia Institute of Technology Systems Realization Laboratory 1 Primal Methods.
Department of Statistics University of Rajshahi, Bangladesh
STROUD Worked examples and exercises are in the text Programme 10: Sequences PROGRAMME 10 SEQUENCES.
Inequality Constraints Lecture 7. Inequality Contraints (I) n A Review of Lagrange Multipliers –As we discussed last time, the first order necessary conditions.
Support Vector Machine: An Introduction. (C) by Yu Hen Hu 2 Linear Hyper-plane Classifier For x in the side of o : w T x + b  0; d = +1; For.
Searching a Linear Subspace Lecture VI. Deriving Subspaces There are several ways to derive the nullspace matrix (or kernel matrix). ◦ The methodology.
Amir Yavariabdi Introduction to the Calculus of Variations and Optical Flow.
D Nagesh Kumar, IISc Water Resources Systems Planning and Management: M2L2 Introduction to Optimization (ii) Constrained and Unconstrained Optimization.
Ch 9.6: Liapunov’s Second Method In Section 9.3 we showed how the stability of a critical point of an almost linear system can usually be determined from.
SEQUENCES A function whose domain is the set of all integers greater than or equal to some integer n 0 is called a sequence. Usually the initial number.
1 Chapter 4 Geometry of Linear Programming  There are strong relationships between the geometrical and algebraic features of LP problems  Convenient.
Chapter 5: Integration Section 5.1 An Area Problem; A Speed-Distance Problem An Area Problem An Area Problem (continued) Upper Sums and Lower Sums Overview.
Chap 3. The simplex method
Outline Unconstrained Optimization Functions of One Variable
Derivatives in Action Chapter 7.
Multivariable optimization with no constraints
Presentation transcript:

Linear & Nonlinear Programming -- Basic Properties of Solutions and Algorithms

2 Outline  First-order Necessary Condition  Examples of Unconstrained Problems  Second-order Conditions  Convex and Concave Functions  Minimization and Maximization of Convex Functions  Global Convergence of Descent Algorithms  Speed of Convergence

3 Introduction  Considering optimization problem of the form where f is a real-valued function and Ω, the feasible set, is a subset of E n. where f is a real-valued function and Ω, the feasible set, is a subset of E n.

4 Weierstras Theorem  If f is continuous and Ω is compact, a solution point of minimization problem exists.

5 Two Kinds of Solution Points  Definition of a relative minimum point –A pointis said to be a relative minimum point of f over Ω if there is an such that f(x) ≥ f(x * ) for all within a distance of x *. If f(x) > f(x * ) for all, x≠x *, within a distance of x *, then x * is said to be a strict relative minimum point of f over.  Definition of a global minimum point –A pointis said to be a global minimum point of f over Ω if f(x) ≥ f(x * ) for all. If f(x) > f(x * ) for all, x≠x *, then x * is said to be a strict global minimum point of f over.

6 Two Kinds of Solution Points (cont’d)  We can achieve relative minimum by using differential calculus or a convergent stepwise procedure.  Global conditions and global solutions can, as a rule, only be found if the problem possesses certain convexity properties that essentially guarantee that any relative minimum is a global minimum.

7 Feasible Directions  To derive necessary conditions satisfied by a relative minimum point x *, the basic idea is to consider movement away from the point in some given direction.  A vector d is a feasible direction at x if there is an such that for all α,.

8 Feasible Directions (cont’d)  Proposition 1 ( first-order necessary conditions) –Let Ω be a subset of E n and let be a function on Ω. If x* is a relative minimum point of f over Ω, then for any that is a feasible direction at x*, we have Proof : Proof :

9 Feasible Directions (cont’d)  Corollary ( unconstrained case ) –Let Ω be a subset of E n and let be a function on Ω. If x* is a relative minimum point of f over Ω and x* is an interior point of Ω, then. Since in this case d can be any direction from x*, and hence for all. This implies. for all. This implies.

10 Example 1 about Feasible Directions  Example 1 ( unconstrained ) minimize minimize There are no constrains, so Ω = E n There are no constrains, so Ω = E n These have the unique solution x 1 =1, x 2 =2, so it is a global minimum of f.

11 Example 2 about Feasible Directions  Example 2 (a constrained case) minimize subject to x 1 ≥ 0, x 2 ≥ 0. since we know that there is a global minimum at x 1 =1/2, x 2 =0, then

12 Example 2 (cont’d) x1x1 x2x2 0 (1/2, 0) 1 3/22 1/2 1 3/2 d = ( d 1, d 2 ) Feasible region

13 Example 3 of Unconstrained Problems  The problem faced by an electric utility when selecting its power-generating facilities. Its power-generating requirements are summarized by a curve, h(x), as shown in Fig.6.2(a), which shows the total hours in a year that a power level of at least x is required for each x. For convenience the curve is normalized so that the upper limit is unity. The power company may meet these requirements by ins- talling generating equipment, such as (1) nuclear or (2) coal-fired, or by purchasing power from a central energy.

14 Example 3 (cont’d)  Associated with type i ( i = 1,2 ) of generating equipment is a yearly unit capital cost b i and a unit operating cost c i. The unit price of power purchased from the grid is c 3. The requirements are satisfied as shown in Fig.6.2(b), where x 1 and x 2 denote the capacities of the nuclear and coal-fired plants, respectively.

15 Example 3 (cont’d) Fig.6.2 Power requirement curve hours required h(x) power (megawatts) 1 (a) hours required h(x) power (megawatts) 1 (b) x1x1 x2x2 coal nuclea r purchas e

16 Example 3 (cont’d)  The total cost is  And the company wishes to minimize the set defined by x 1 ≥ 0, x 2 ≥ 0, x 1 + x 2 ≤ 1 x 1 ≥ 0, x 2 ≥ 0, x 1 + x 2 ≤ 1

17 Example 3 (cont’d)  Assume that the solution is interior to the constraints, by setting the partial derivatives equal to zero, we obtain the two equations which represent the necessary conditions  In addition, If x 1 =0, then equality (1) relax to ≥ 0 If x 2 =0, then equality (2) relax to ≥ 0

18 Second-order Conditions  Proposition 1 ( second-order necessary conditions ) –Let Ω be a subset of E n and let be a function on Ω. if x* is a relative minimum point of f over Ω, then for any which is a feasible direction at x* we have i) ii) if, then Proof : Proof : The first condition is just propotion1, and the second condition applies only if.

19 Proposition 1 (cont’d)  Proof (cont’d) :

20 Example 1 about Proposition 1  For the same problem as Example 2 of Section 6.1, we have d = (d 1, d 2 ) Example 2 of Section 6.1Example 2 of Section 6.1  Thus condition (ii) of Proposition 1 applies only if d 2 = 0. In that case we have so condition (ii) is satisfied. so condition (ii) is satisfied.

21 Proposition 2  Proposition 2 (unconstrained case) –Let x* be an interior point of the set Ω, and suppose x* is a relative minimum point over Ω of the function. Then i) ii) for all d,. –It means that F(x*), simplified notation of, is positive semi-definite.

22 Example 2 about Proposition 2  Consider the problem –If we assume the solution is in the interior of the feasible set, that is, if x 1 > 0, x 2 > 0, then the first-order necessary conditions are

23 Example 2 (cont’d)  Boundary solution is x 1 = x 2 = 0  Another solution at x 1 = 6, x 2 = 9 –If we fixed x 1 at x 1 = 6, then the relative minimum with respect to x 2 at x 2 = 9. –Conversely, with x 2 fixed at x 2 = 9, the objective attains a relative minimum w.r.t. x 1 at x 1 = 6. –Despite this fact, the point x 1 = 6, x 2 = 9 is not a relative minimum point, because the Hessian matrix F(x*) x1 = 6, x2 = 9 is not a positive semi-definite since its determinant is negative.

24 Sufficient Conditions for a Relative Minimum  We give here the conditions that apply only to unconstrained problems, or to problems where the minimum point is interior to the feasible solution.  Since the corresponding conditions for problems where the minimum is achieved on a boundary point of the feasible set are a good deal more difficult and of marginal practical or theoretical value.  A more general result, applicable to problems with functional constrains, is given in Chapter10.

25 Proposition 3  Proposition 3 (2-order sufficient conditions-unconstrained case) –Let be a function defined on a region in which the point x* is an interior point. Suppose in addition that i) ii) is positive definite Then x* is a strict relative minimum point of f. Proof : since is positive definite, there is an a > 0 such that since is positive definite, there is an a > 0 such that for all d,. Thus by Taylor’s Theorem for all d,. Thus by Taylor’s Theorem

26 Convex Functions  Definition –A function f defined on a convex set Ω is said to be convex. If, for every and every α, 0 ≤ α ≤ 1, there holds If, for every α, 0 < α < 1, and x 1 ≠ x 2, there holds then f is said to be strictly convex.

27 Concave Functions  Definition –A function g defined on a convex set Ω is said to be concave if the function f = -g is convex. The function g is strictly concave if -g is strict convex.

28 Graphs of Strict Convex Function f x1x1 x2x2 α x 1 +(1-α) x 2

29 Graphs of Convex Function f x1x1 x2x2 α x 1 +(1-α) x 2

30 Graphs of Concave Function f x1x1 x2x2 α x 1 +(1-α) x 2

31 Graph of Neither Convex or Concave f x1x1 x2x2

32 Combinations of Convex Functions  Proposition 1 –Let f 1 and f 2 be convex function on the convex set Ω. Then the f 1 + f 2 is convex on Ω.  Proposition 2 –Let f be a convex function over the convex set Ω. Then a f is convex for any a ≥ 0.

33 Combinations (cont’d)  Through the above two propositions it follows that a positive combination a 1 f 1 + a 1 f 2 + …a m f m of is again convex.

34 Convex Inequality Constrains  Proposition 3 –Let f be a convex function on a convex set Ω. The set is a convex for every real number c.

35 Proof –Let, then and for 0 <α< 1, Thus

36 Properties of Differentiable Convex Functions  Proposition 4 –Let, then f is convex over a convex set Ω if and only if for all

37 Recall  The original definition essentially states that linear interpolation between two points overestimates the function,  while here stating that linear approximation based on the local derivative underestimates the function.

38 Recall (cont’d) f is a convex function between two points f x1x1 x2x2 α x 1 +(1-α) x 2

39 Recall (cont’d) f(y) x y

40 Two Continuously Differentiable Functions  Proposition 5 –Let, then f is convex over a convex set Ω containing an interior point if only if the Hessian matrix F of f is positive semi-definite through Ω.

41 Proof  By Taylor’s theorem we have for some α, 0 ≤ α ≤ 1. if the Hessian is everywhere positive semi-definite, we have

42 Minimization and Maximization of Convex Functions  Theorem 1 –Let f be a convex function defined on the convex set Ω, then the setwhere f achieves its minimum is convex, and any relative minimum of f is a global minimum.  Proof (contradiction)

43 Minimization and Maximization of Convex Functions (cont’d)  Theorem 2 –Let be a convex on the convex set Ω. If there is a point such that, for all then x* is a global minimum point of f over Ω.  Proof

44 Minimization and Maximization of Convex Functions (cont’d)  Theorem 3 –Let f be a convex function defined on the bounded, closed convex set Ω. If f has a maximum over Ω, then it is achieved at an extreme point of Ω.

45 Global Convergence of Descent Algorithms  A good portion of the remainder of this book is devoted to presentation and analysis of various algorithms designed to solve nonlinear programming problems. However, they have the common heritage of all being iterative descent algorithms.  Iterative –The algorithm generated a series of points, each point being calculated on the basis of the points preceding it.  Descent –As each new point is generated by the algorithm the corresponding value of some function (evaluated at the most recent point) decreases in values.

46 Global Convergence of Descent Algorithms (cont’d)  Globally convergent –If for arbitrary starting points the algorithm is guaranteed to generate a sequence of points converging to a solution, then the algorithm is said to be globally convergent.

47 Algorithm and Algorithmic Map  We formally define an algorithm A as a mapping taking points in a space X into other points in X, then the generated sequence { x k } defined by –With this intuitive idea of an algorithm in mind, we now generalize the concept somewhat so as to provide greater flexibility in our analysis.  Definition –An algorithm A is a mapping defined on a space X that assigns to every point a subset of X.

48 Mappings  Given the algorithm yields A(x k ) which is a subset of X. From this subset an arbitrary element x k+1 is selected. In this way, given an initial point x 0, the algorithm generates sequences through the iteration  The most important aspect of the definition is that the mapping A is a point-to-set mapping of X.

49 Example 1  Suppose for x on the real line we define so that A(x ) is an interval of the real line. Starting at x 0 = 100, each of the sequences below might be generated from iterative application of this algorithm. 100, 50, 25, 12, -6, -2, 1, 1/2,… 100, -40, 20, -5, -2, 1, 1/4, 1/8,… 100, 10, 1/16, 1/100, -1/1000,…

50 Descent  Definition –Let be a given solution set and let A be an algorithm on X. A continuous real-valued functions Z on X is said to be a descent function for and A if it satisfies

51 Closed Mapping  Definition –A point-to-set mapping A from X to Y is said to be closed at if the assumptions The point-to-set map A is said to be closed on X if it is closed at each point of X.

52 Closed Mapping in Different Space  Many complex algorithms are regards as the composition of two or more simple point-to-point mappings.  Definition –Let be point-to-set mappings. The composite mapping C = BA is defined as the point- to-set mapping The definition is illustrated in Fig. 6.6

53 Fig. 6.6 A B C X Y Z A(x) B(x) C(x) x y

54 Corollaries of Closed Mapping  Corollary 1  Corollary 2

55 Global Convergence Theorem  The Global Convergence Theorem is used to establish convergence for the following situation.  There is a solution set. Points are generated according to the algorithm, and each new point always strictly decreases a descent function Z unless the solution set is reached.  For example, in nonlinear programming, the solution set may be the set of minimum points (perhaps only one point), and the descent function may be the objective function itself.

56 Global Convergence Theorem (cont’d)  Let A be an algorithm on X, and suppose that, given x 0 the sequence is generated satisfying let a solution set be given, and suppose Then the limit of any convergence subsequence of {x k } is a solution

57 Global Convergence Theorem (cont’d)  Corollary –If under the conditions of the Global Convergence Theorem consists of a single point, then the sequence {x k } converges to.

58 Examples to Illustrate Conditions  Examples 4 –On the real line consider the point-to-point algorithm and solution set,descent function The condition (iii) does not holds since A is not closed at x = 1, so, the limit of any convergent subsequence of {x k } is not a solution.

59 Examples (cont’d)  Example 5 –On the real line X consider the solution set, the descent function Z(x) = e -x, and the algorithm A(x) = x + 1. All the conditions of the convergence theorem holds except (i) since the sequence generated from any starting point x 0 diverges to infinity. It means S is not a compact set.

Order of convergence  Definition –Let the sequence converge to the limit r*. The order of convergence of {r k } is defined as the supremum of the nonnegative numbers p satisfying –Larger values of the order p imply faster convergence.

61 Order of convergence (cont’d) –If the sequence has order p and the limit exists, then asymptotically we have

62 Examples  Example 1 –The sequence with r k = a k where 0 < a < 1 Solution –It converges to zero with order unity, since

63 Examples (cont’d)  Example 2 –The sequence with for 0 < a < 1 Solution –It converges to zero with order two, since

64 Linear convergence  Most algorithm discussed in this book have an order of convergence equal to unity.  Definition –If the sequence {r k } converges to r * in such way that the sequence is said to converge linearly to r * with convergence ratio

65 Linear convergence (cont’d)  The linearly convergence is sometimes referred to as geometric convergence.  Since one with convergence ratio can be said to have a tail that converges at least as fast as the geometric sequence for some constant c.  The smaller the ratio the faster the rate.  The ultimate case where is referred to as superlinear convergence.

66 Examples  Example 3 –The sequence with r k = 1/k Solution –It converges to zero. The convergence is of order one but it is not linear, since is not less than one.

67 Examples (cont’d)  Example 4 –The sequence with r k = (1/k) k Solution –The sequence is of order unity, since for p > 1. for p > 1.However, and hence this is superlinear convergence.

68 Average rates  All the definitions given above can be referred to as step-wise concepts of convergence, since they define bounds on the progress made by going a single step : from k to k+1.  Another approach is to define concepts related to the average progress per step over a large number of steps.

69 Average rates (cont’d)  Definition –Let the sequence {r k } converge to r *. The average order of convergence is the infimum of the numbers p > 1 such that The order is infinity if the equality holds for no p > 1

70 Examples  Example 5 –For the sequence, 0 < a < 1 Solution –for p = 2, we have, for p > 2, we have Thus the average order is two

71 Examples (cont’d)  Example 6 –For the sequence r k = a k, with 0 < a < 1 Solution –for p > 1,we have Thus the average order is unity.

72 Convergence ratio by average method  The most important case is that of unity order.  We define the average convergence ratio as  Thus for the geometric sequence r k = ca k, for 0 < a < 1 the average convergence ratio is a.