Escaping Local Optima. Where are we? Optimization methods Complete solutions Partial solutions Exhaustive search Hill climbing Exhaustive search Hill.

Slides:



Advertisements
Similar presentations
Local optimization technique G.Anuradha. Introduction The evaluation function defines a quality measure score landscape/response surface/fitness landscape.
Advertisements

Local Search Algorithms
G5BAIM Artificial Intelligence Methods
Neural and Evolutionary Computing - Lecture 4 1 Random Search Algorithms. Simulated Annealing Motivation Simple Random Search Algorithms Simulated Annealing.
Tabu Search Strategy Hachemi Bennaceur 5/1/ iroboapp project, 2013.
Search by partial solutions. Where are we? Optimization methods Complete solutions Partial solutions Exhaustive search Hill climbing Random restart General.
CPSC 322, Lecture 16Slide 1 Stochastic Local Search Variants Computer Science cpsc322, Lecture 16 (Textbook Chpt 4.8) February, 9, 2009.
Spie98-1 Evolutionary Algorithms, Simulated Annealing, and Tabu Search: A Comparative Study H. Youssef, S. M. Sait, H. Adiche
Tabu Search for Model Selection in Multiple Regression Zvi Drezner California State University Fullerton.
Local search algorithms
Local search algorithms
Two types of search problems
Heuristic Optimization Athens 2004 Department of Architecture and Technology Universidad Politécnica de Madrid Víctor Robles
MAE 552 – Heuristic Optimization Lecture 27 April 3, 2002
Ryan Kinworthy 2/26/20031 Chapter 7- Local Search part 1 Ryan Kinworthy CSCE Advanced Constraint Processing.
Optimization via Search CPSC 315 – Programming Studio Spring 2009 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
Local Search and Stochastic Algorithms Solution tutorial 4.
Introduction to Artificial Intelligence Local Search (updated 4/30/2006) Henry Kautz.
MAE 552 – Heuristic Optimization Lecture 4 January 30, 2002.
Review Best-first search uses an evaluation function f(n) to select the next node for expansion. Greedy best-first search uses f(n) = h(n). Greedy best.
Simulated Annealing Van Laarhoven, Aarts Version 1, October 2000.
1 IOE/MFG 543 Chapter 14: General purpose procedures for scheduling in practice Section 14.4: Local search (Simulated annealing and tabu search)
MAE 552 – Heuristic Optimization Lecture 5 February 1, 2002.
Introduction to Simulated Annealing 22c:145 Simulated Annealing  Motivated by the physical annealing process  Material is heated and slowly cooled.
Optimization via Search CPSC 315 – Programming Studio Spring 2008 Project 2, Lecture 4 Adapted from slides of Yoonsuck Choe.
By Rohit Ray ESE 251.  Most minimization (maximization) strategies work to find the nearest local minimum  Trapped at local minimums (maxima)  Standard.
Metaheuristics The idea: search the solution space directly. No math models, only a set of algorithmic steps, iterative method. Find a feasible solution.
Elements of the Heuristic Approach
Informed Search Methods
D Goforth - COSC 4117, fall Note to 4 th year students  students interested in doing masters degree and those who intend to apply for OGS/NSERC.
Local Search and Optimization
1 Local search and optimization Local search= use single current state and move to neighboring states. Advantages: –Use very little memory –Find often.
Local Search. m = 3h - 8 linear fitness function – optimum at domain boundary.
An Introduction to Artificial Life Lecture 4b: Informed Search and Exploration Ramin Halavati In which we see how information.
Local Search Algorithms This lecture topic Chapter Next lecture topic Chapter 5 (Please read lecture topic material before and after each lecture.
Search Methods An Annotated Overview Edward Tsang.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
1 Simulated Annealing Contents 1. Basic Concepts 2. Algorithm 3. Practical considerations.
Escaping Local Optima. Where are we? Optimization methods Complete solutions Partial solutions Exhaustive search Hill climbing Random restart General.
Simulated Annealing.
Computer Science CPSC 322 Lecture 16 CSP: wrap-up Planning: Intro (Ch 8.1) Slide 1.
Thursday, May 9 Heuristic Search: methods for solving difficult optimization problems Handouts: Lecture Notes See the introduction to the paper.
Applications of Dynamic Programming and Heuristics to the Traveling Salesman Problem ERIC SALMON & JOSEPH SEWELL.
Local Search Pat Riddle 2012 Semester 2 Patricia J Riddle Adapted from slides by Stuart Russell,
G5BAIM Artificial Intelligence Methods
Single-solution based metaheuristics. Outline Local Search Simulated annealing Tabu search …
Introduction to Simulated Annealing Study Guide for ES205 Xiaocang Lin & Yu-Chi Ho August 22, 2000.
Simulated Annealing. Difficulty in Searching Global Optima starting point descend direction local minima global minima barrier to local search.
Local search algorithms In many optimization problems, the state space is the space of all possible complete solutions We have an objective function that.
Optimization Problems
Lecture 6 – Local Search Dr. Muhammad Adnan Hashmi 1 24 February 2016.
1 Simulated Annealing Contents 1. Basic Concepts 2. Algorithm 3. Practical considerations.
CPSC 322, Lecture 16Slide 1 Stochastic Local Search Variants Computer Science cpsc322, Lecture 16 (Textbook Chpt 4.8) Oct, 11, 2013.
Constraints Satisfaction Edmondo Trentin, DIISM. Constraint Satisfaction Problems: Local Search In many optimization problems, the path to the goal is.
Local search algorithms In many optimization problems, the path to the goal is irrelevant; the goal state itself is the solution State space = set of "complete"
1 Intro to AI Local Search. 2 Intro to AI Local search and optimization Local search: –use single current state & move to neighboring states Idea: –start.
School of Computer Science & Engineering
CSCI 4310 Lecture 10: Local Search Algorithms
Department of Computer Science
Heuristic Optimization Methods
Van Laarhoven, Aarts Version 1, October 2000
Local Search Strategies: From N-Queens to Walksat
Artificial Intelligence (CS 370D)
Subject Name: Operation Research Subject Code: 10CS661 Prepared By:Mrs
Hill-climbing Search Goal: Optimizing an objective function.
Metaheuristic methods and their applications. Optimization Problems Strategies for Solving NP-hard Optimization Problems What is a Metaheuristic Method?
School of Computer Science & Engineering
Artificial Intelligence
Local Search Algorithms
Simulated Annealing & Boltzmann Machines
Presentation transcript:

Escaping Local Optima

Where are we? Optimization methods Complete solutions Partial solutions Exhaustive search Hill climbing Exhaustive search Hill climbing Exhaustive search Branch and bound Greedy Best first A* Divide and Conquer Dynamic programming Exhaustive search Branch and bound Greedy Best first A* Divide and Conquer Dynamic programming

Where are we going? Optimization methods Complete solutions Partial solutions Exhaustive search Branch and bound Greedy Best first A* Divide and Conquer Dynamic programming Exhaustive search Branch and bound Greedy Best first A* Divide and Conquer Dynamic programming Improved methods based on complete solutions Exhaustive search Hill climbing Exhaustive search Hill climbing

Escaping local optima Many strategies including Simulated annealing Tabu search and many others… some examples -->

(0) Basic Hill climbing determine initial solution s while s is not a local optimum choose s’ in N(s) such that f(s’)>f(s) s = s’ return s

(1) Randomized Hill climbing determine initial solution s; bestS = s while termination condition not satisfied with probability p choose neighbour s’ at random (uniform) else choose s’ with f(s’) > f(s) //climb if possible or s’ with max (f(s’)) over N(s) s = s’; if (f(s) > f(bestS)) bestS = s return bestS

(2) Variable Neighbourhood determine initial solution s i = 1 repeat choose neighbour s’ in N i (s) with max(f(s’)) if ((f(s’) > f(s)) s = s’ i = 1// restart in first neighbourhood else i = i+1// go to next neighbourhood until i > iMax return s

Stochastic local search  many other important algorithms address the problem of avoiding the trap of local optima (possible source of project topics)  M&F focus on two only simulated annealing tabu search

Simulated annealing  metaphor: slow cooling of liquid metals to allow crystal structure to align properly  “temperature” T is slowly lowered to reduce random movement of solution s in solution space

Simulated Annealing determine initial solution s; bestS = s T = T 0 while termination condition not satisfied choose s’ in N(s) probabilistically if (s’ is “acceptable”) // function of T s = s’ if (f(s) > f(sBest)) bestS = s update T return bestS

Accepting a new solution - acceptance more likely if f(s’) > f(s) - as execution proceeds, probability of acceptance of s’ with f(s’) < f(s) decreases (becomes more like hillclimbing) determine initial solution s; bestS = s T = T 0 while termination condition not satisfied choose s’ in N(s) probabilistically if (s’ is “acceptable”) // function of T s = s’ if (f(s) > f(sBest)) bestS = s update T return bestS determine initial solution s; bestS = s T = T 0 while termination condition not satisfied choose s’ in N(s) probabilistically if (s’ is “acceptable”) // function of T s = s’ if (f(s) > f(sBest)) bestS = s update T return bestS

the acceptance function T evolves *sometimes p=1 when f(s’)-f(s)> 0

Simulated annealing with SAT algorithm p.123 SA-SAT propositions:P 1,… P n expression:F = D 1 D 2 …D k where clause D i is a disjunction of propositions and negative props e.g.,P x  ~P y  P z  ~P w fitness function:number of true clauses

Inner iteration assign random truth setTFFT repeat for i=1 to 4 flip truth of prop iFFFT evaluateFTFT decide to keep (or not)FFTT changed valueFFTF reduce TFFTT

Tabu search (taboo) always looks for best solution but some choices (neighbours) are ineligible (tabu) ineligibility is based on recent moves: once a neighbour edge is used, it is tabu for a few iterations search does not stop at local optimum

Symmetric TSP example set of 9 cities {A,B,C,D,E,F,G,H,I} neighbour definition based on 2-opt* (27 neighbours) current sequence: B - D - A - I - H - F - E - C - G - B move to 2-opt neighbour B - E - F - H - I - A - D - C - G - B edges B-E and D-C are now tabu i.e., next 2-opt swap cannot involve these edges *example in book uses 2-swap, p 131

TSP example, algorithm p 133 how long will an edge be tabu?3 iterations how to track and restore eligibility? data structure to store tabu status of 9*8/2 = 36 edges B - D - A - I - H - F - E - C - G - B recency-based memory ABCDEFGH I H G F00000 E0030 D230 C00 B0

procedure tabu search begin tries <- 0 repeat generate a tour count <- 0 repeat identify a set T of 2-opt moves select best admissible move from T make appropriate 2-opt update tabu list and other vars if new tour is best-so-far for a given tries update local best tour information count <- count + 1 until count == ITER tries <- tries + 1 if best-so-far for given tries is best-so-far (for all ‘tries’) update global best information until tries == MAX-TRIES end

applying 2-opt with tabu  from the table, some edges are tabu: B - D - A - I - H - F - E - C - G - B  2-opt can only consider:  AI and FE  AI and CG  FE and CG ABCDEFGH I H G F00000 E0030 D230 C00 B0

importance of parameters  once algorithm is designed, it must be “tuned” to the problem selecting fitness function and neighbourhood definition setting values for parameters  this is usually done experimentally

procedure tabu search begin tries <- 0 repeat generate a tour count <- 0 repeat identify a set T of 2-opt moves select best admissible move from T make appropriate 2-opt update tabu list and other vars if new tour is best-so-far for a given tries update local best tour information count <- count + 1 until count == ITER tries <- tries + 1 if best-so-far for given tour is best-so-far for all tries update global best information until tries == MAX-TRIES end