Presentation is loading. Please wait.

Presentation is loading. Please wait.

PARTICLE SWARM OPTIMISATION (PSO) Perry Brown Alexander Mathews Image:

Similar presentations


Presentation on theme: "PARTICLE SWARM OPTIMISATION (PSO) Perry Brown Alexander Mathews Image:"— Presentation transcript:

1 PARTICLE SWARM OPTIMISATION (PSO) Perry Brown Alexander Mathews Image: http://www.cs264.org/2009/projects/web/Ding_Yiyang/ding-robb/pso.jpg

2 Introduction Concept first introduced by Kennedy and Eberhart (1995) Original idea was to develop a realistic visual simulation of bird flock behaviour Simulation was then modified to include a point in the environment that attracted the virtual bird agents Potential for optimisation applications then became apparent

3 The natural metaphor A flock of birds (or school of fish, etc.) searching for food Their objective is to efficiently find the best source of food Nature-based theory underlying PSO: The advantage of sharing information within a group outweighs the disadvantage of having to share the reward Image: http://www.nerjarob.com/nature/wp- content/uploads/Flock-of-pigeons.jpg

4 Terminology The “particles” in PSO have no mass or volume (essentially they are just points in space), but they do have acceleration and velocity Behaviour of groups in the developed algorithm ended up looking more like a swarm than a flock Hence the name Particle Swarm Optimisation

5 Swarm intelligence Millonas’ five basic principles of swarm intelligence:  Proximity: agents can perform basic space and time calculations  Quality: agents can respond to environmental conditions  Diverse response: population can exhibit a wide range of behaviour  Stability: behaviour of population does not necessarily change every time the environment does  Adaptability: behaviour of population must be able to change when necessary to adapt to environment A PSO swarm satisfies all of the above conditions

6 Population and environment Multidimensional search space Each point in the search space has some value associated with it, goal is to find the “best” value Numerous agent particles navigating the search space Each agent has the following properties:  a current position within the search space  a velocity vector Additionally, each agent knows the following information:  the best value it has found so far (pbest) and its location  the best value any member of the population has found so far (gbest) and its location

7 Kennedy and Eberhart’s (1995) refined algorithm Some number of agent particles are initialised with individual positions and velocities (often just done randomly) The following steps are then performed iteratively:  The position of each agent is updated according to its current velocity: new position = old position + velocity  The value at each agent’s new position is checked, with pbest and gbest information updated if necessary  Each component of each agent’s velocity vector is then adjusted as a function of the differences between its current location and both the pbest and gbest locations, each weighted by a random variable: new velocity = old velocity + 2 * rand1 * (pbest location - current location) + 2 * rand2 * (gbest location - current location) where rand1 and rand2 are random numbers between 0 and 1. (Multiplying by the constant 2 causes particles to “overshoot” their target about half of the time, resulting in further exploration.)

8 A (partial) example in two dimensions 0123456789 0321 1311 221221 313432 42453111 513232121 61311132 71321121 8142111 9121 pbests Blue: 0 Green: 0 Red: 0 gbest: 0 (dots indicate agents, yellow star indicates the global optimum)

9 Begin with random velocities 0123456789 0321 1311 221221 313432 42453111 513232121 61311132 71321121 8142111 9121 pbests Blue: 0 Green: 0 Red: 0 gbest: 0

10 Update particle positions 0123456789 0321 1311 221221 313432 42453111 513232121 61311132 71321121 8142111 9121 pbests Blue: 1 at (6, 2) Green: 0 Red: 2 at (8, 7) gbest: 2 at (8, 7)

11 Update particle velocities 0123456789 0321 1311 221221 313432 42453111 513232121 61311132 71321121 8142111 9121 pbests Blue: 1 at (6, 2) Green: 0 Red: 2 at (8, 7) gbest: 2 at (8, 7) For example, Blue’s velocity in the horizontal dimension calculated by: velocity = 1 + 2 * rand() * (6 – 6) + 2 * rand() * (8 – 6)

12 Update particle positions again and repeat… 0123456789 0321 1311 221221 313432 42453111 513232121 61311132 71321121 8142111 9121 pbests Blue: 3 at (8, 6) Green: 1 at (4, 1) Red: 2 at (8, 7) gbest: 3 at (8, 6)

13 Algorithm termination The solution to the optimisation problem is (obviously) derived from gbest Possible termination conditions that might be used:  Solution exceeds some quality threshold  Average velocity of agents falls below some threshold (agents may never become completely stationary)  A certain number of iterations is completed

14 An example visualisation http://www.youtube.com/watch?v=_bzRHqmpwvo Velocities represented by trailing lines After some individual exploration, particles all converge on global optimum Particles can be seen oscillating about the optimum

15 Reference Kennedy, J.; Eberhart, R.;, "Particle swarm optimization," Neural Networks, 1995. Proceedings., IEEE International Conference on, vol.4, no., pp.1942-1948 vol.4, Nov/Dec 1995 URL: http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=48896 8&isnumber=10434http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=48896 8&isnumber=10434


Download ppt "PARTICLE SWARM OPTIMISATION (PSO) Perry Brown Alexander Mathews Image:"

Similar presentations


Ads by Google