Presentation is loading. Please wait.

Presentation is loading. Please wait.

Constraint Optimization We are interested in the general non-linear programming problem like the following Find x which optimizes f(x) subject to gi(x)

Similar presentations


Presentation on theme: "Constraint Optimization We are interested in the general non-linear programming problem like the following Find x which optimizes f(x) subject to gi(x)"— Presentation transcript:

1 Constraint Optimization We are interested in the general non-linear programming problem like the following Find x which optimizes f(x) subject to gi(x)  0, i=1, ,n hj(x)=0, j=1, ,p

2 How to solve Penalty functions Special representations and operators Repair algorithms Separation of objectives and constraints Hybrid methods

3 Penalty functions The most common approach to handle constraints (particularly, inequality constraints) is to use penalties. The basic idea is to transform a constrained-optimization problem into an unconstrained one by adding/substracting a certain value to/from the objective function based on the amount of constraint violation present in a certain solution.

4 General formulation of the penalty function The general formulation of the penalty function is  (x)=f(x)  (  ri  Gi+  cj  Lj) where  (x) is the new objective function to be optimized, Gi and Lj are functions of the constraints gi(x) and hj(x), respectively, and ri and cj are positive constants normally called “penalty factors” The most common form of Gi and Lj is Gi=max(0,gi(x))  Lj=|hj(x)|  where  and  are normally 1 or 2

5 Minimum penalty rule The penalty should be kept as low as possible, just above the limit below which infeasible solutions are optimal. If the penalty is too high and the optimum lies at the boundary of the feasible region, the GA will be pushed inside the feasible region very quickly and will not be able to move back towards the boundary with the infeasible region. If there are several disjoint feasible regions in the search space, the GA would tend to move to one of them and would not be able to move to a different feasible region. If the penalty is too low, a lot of the search time will be spent exploring the infeasible region because the penalty will be negligible with respect to the objective functions.

6 Static Penalties The penalty factors remain constant during the entire evolutionary process. Fitness(x)=f(x)  (Ri  max[0,gi(x)] 2 )

7 Dynamic penalties Current generation number is involved in the computation of the corresponding penalty factors Fitness(x)=f(x)  ((Ri  t)   max[0,gi(x)] 2 )

8 Annealing penalties Penalty factors are changed once in many generations (after the algorithm has been trapped in a local optima) Fitness(x)=f(x)  max[0,gi(x)] 2 /  where  is the cooling schedule.

9 Adaptive penalties Use a penalty function which takes a feedback from the search process. Fitness(x)=f(x)  (t)  max[0,gi(x)] 2 where (t) is updated at every generation t in the following way If the best individual in the last k generations was always feasible: (t+1)= (t)/ ,  >1 If the best individual in the last k generations was never feasible: (t+1)=  (t),  >1 Otherwise, there are some feasible and infeasible individuals tied as best in the population, the penalty does not change.

10 Segregated GA Use two penalty parameters instead of one. These two values aim at achieving a balance between heavy and moderate penalties. A population of size 2m is generated. These individuals are divided into two groups. Each group use one penalty parameter. Choose m best individuals from these two groups to become parents for the next generation. These m parents produce m offspring. The 2m individuals are combined together as a new generation.

11 Death penalty The rejection of infeasible individuals is probably the easiest way to handle constraints.

12 Special Representations and operators Solve a certain problem for which a generic representation scheme might not be appropriate To simplify the shape of the search space It is always difficult to locate at least a single feasible solution.

13 Linear constraints Eliminate equality constraints together with an equal number of problem variables The search space is a convex sets. Crossover: linear combinations of individuals Mutation: Re-define the domain.

14 Linear constraints: example Optimize f(x1,x2,x3,x4,x5,x6) 2x1+x2+x3=6 x3+x5-3x6=10 x1+4x4=3 x2+x5  120 -40  x1  20 50  x2  75 0  x3  10 5  x4  15 0  x5  20 -5  x6  5

15 Eliminate equality constraints x1=3-4x4 x2=-10+8x4+x5-3x6 x3=10-x5+3x6

16 New problem Optimize g(x4,x5,x6)=f(3-4x4,-10+8x4+x5-3x6,10- x5+3x6,x4,x5,x6) -10+8x4+2x5-3x6  120 (x2+x5  120) -40  3-4x4  20 (-40  x1  20) 50  -10+8x4+x5-3x6  75 (50  x2  75) 0  10-x5+3x6  10 (0  x3  10) 5  x4  15 0  x5  20 -5  x6  5 Combine 2nd and 5th: 5  x4  10.75

17 Mutation If (x4,x5,x6)=(10,8,2), then x4  [7.25,10.375] x5  [6,11] x6  [1,2.666]

18 Heuristic crossover Suppose f(x2) is better than f(x1) x3=r  (x2-x1)+x2 r  [0,1] Check if x3 satisfy all the constraints

19 Decoder A chromosome “gives instructions” on how to build a feasible solution. For each feasible solution there must be a decoded solution Each decoded solution must correspond to a feasible solution The transformation is computationally fast and it has locality feature in the sense that small changes in the decoded solution result in small changes in the solution itself

20 TSP and ordinal expression The ith gene belongs to the region [1,n-i+1]. It determines which city will be chosen from the city list.

21 Ordinal expression City list (1 2 3 4 5 6 7 8 9) Chromosome ( 1 1 2 1 4 1 3 1 1) Then the path is (1 2 4 3 8 5 9 6 7)

22 Ordinal expression City list (1 2 3 4 5 6 7 8 9) Chromosome ( 1 1 2 1 4 1 3 1 1) Then the path is (1) City list (2 3 4 5 6 7 8 9) Chromosome ( 1 1 2 1 4 1 3 1 1) Then the path is (1 2) City list (3 4 5 6 7 8 9) Chromosome ( 1 1 2 1 4 1 3 1 1) Then the path is (1 2 4) City list (3 5 6 7 8 9) Chromosome ( 1 1 2 1 4 1 3 1 1) Then the path is (1 2 4 3) City list (5 6 7 8 9) Chromosome ( 1 1 2 1 4 1 3 1 1) Then the path is (1 2 4 3 8) …...

23 Crossover The traditional one-point crossover can be used p1=(1 1 2 1. 4 1 3 1 1) p2=(5 1 5 5. 5 3 3 2 1) which correspond to the paths (1 2 4 3 8 5 9 6 7) (5 1 7 8 9 4 6 3 2)

24 Crossover After crossover c1=(1 1 2 1 5 3 3 2 1) c2=(5 1 5 5 4 1 3 1 1) The new paths are (1 2 4 3 9 7 8 6 5) (5 1 7 8 6 2 9 3 4)

25 Repair algorithms Repair an infeasible individual, i.e., to make feasible an infeasible individual. Such a repaired version can be used either for evaluation only, or it can also replace (with some probability) the original individual in the population.

26 Basic rule There are no standard heuristics for the design of repair algorithms The success of this approach relies mainly on the ability of the user to come up with such a heuristics. It is possible to use a greedy algorithm (i.e., an optimization algorithm that proceeds through a series of alternatives by making the best decision, as computed locally, at each point in the series), a random algorithm or any other heuristic which would guide the repair process.

27 Repair by random evolution The main idea is to use random evolutionary search combined with a mathematical programming technique for unconstrained optimization. Whenever a solution is not feasible, the following constraint functional is minimized C(x)=  hj 2 (x)-  gj(x) c1={i=1, ,n| |hi(x)|>  } c2={j=1, ,q| gj(x)<0}

28 Separation of constraints and objectives Handle constraints and objectives separately Superiority of feasible points Behavioral memory

29 Superiority of feasible points Evaluations of feasible solutions are mapped into a better interval, and infeasible solution into a worst interval Modified tournament selection A feasible solution is always better than an infeasible one Between two feasible solutions, the one having a better objective function value is preferred. Between two infeasible solutions, the one having smaller constraints violation is preferred

30 Behavioral memory Constraints are handled in a particular order Start with a random population of individuals Set j=1 (j is the constraint counter) Evolve this population to minimize the violation of the jth constraint, until a given percentage of the population is feasible for this constraint. Points that do not satisfy at least one of the 1st, 2nd, , (j-1)th constraints are eliminated from the population. j=j+1. If j<=m repeat the former step. Optimize the objective function and reject infeasible individuals.

31 Hybrid methods In this section we are considering methods that are coupled with another technique (normally a numerical optimization approach) to handle constraints.

32 Fuzzy logic Replace constraints of the form gi(x)  bi by a set of fuzzy functions ci(x) if gi(x)  bi, ci(x)=1 if bi<gi(x)  bi+s, ci(x)=(exp(-p((x-bi)/s)^2)-exp(-p))/(1-exp(-p)) if gi(x)>bi+s, ci(x)=0 This method allows a higher degree of tolerance if gi(x) is greater than bi but close to bi. The tolerance decrease rapidly when the error increase. fitness(x)=f(x)  min(c1(x), ,cm(x))

33 Reliability of Communication System

34 R(k): valid probability of component k Q(k): invalid probability of component k Rs: valid probability of the system Qs: invalid probability of the system Qs=(Q(1)Q(4)) 2 R(3)+(Q(2)+R(2)Q(1)Q(4)) 2 Q(3) C(k)=K(k)R(k)^a(k): cost of component k Cs=2C(1)+2C(2)+C(3)+2C(4): cost of the system

35 Minimum cost problem Min Cs s.t. Rmin  Rs Rmin(k)  R(k)  1.0 We use penalty function Min Cs+  max[0, Rmin-Rs] s.t. Rmin(k)  R(k)  1.0

36 Maximum reliability min Qs s.t. Cs  Cmax Rmin(k)  R(k)  1.0 We use variable replacement: x(k)=R(k)^a(k) Then the constraints are 2K(1)X(1)+2K(2)X(2)+K(3)X(3)+2K(4)X(4)  Cmax Rmin(k)^a(k)  X(k)  1.0


Download ppt "Constraint Optimization We are interested in the general non-linear programming problem like the following Find x which optimizes f(x) subject to gi(x)"

Similar presentations


Ads by Google