Local Search Algorithms Chapter 4. Outline Hill-climbing search Simulated annealing search Local beam search Genetic algorithms Ant Colony Optimization.

Slides:



Advertisements
Similar presentations
Informed search algorithms
Advertisements

Informed search algorithms
CMSC 671 Fall 2005 Class #5 – Thursday, September 15.
Local Search Algorithms
Informed search algorithms
Search I: Chapter 3 Aim: achieving generality Q: how to formulate a problem as a search problem?
STRATEGI PENCARIAN DENGAN INFORMASI (INFORMED SEARCH STRATEGY)
Strategi Pencarian dengan Informasi (Informed Search Strategy)
LOCAL SEARCH AND CONTINUOUS SEARCH. Local search algorithms  In many optimization problems, the path to the goal is irrelevant ; the goal state itself.
Problem Solving by Searching
CSC344: AI for Games Lecture 5 Advanced heuristic search Patrick Olivier
Local search algorithms
Local search algorithms
Two types of search problems
1 Chapter 5 Advanced Search. 2 Chapter 5 Contents l Constraint satisfaction problems l Heuristic repair l The eight queens problem l Combinatorial optimization.
CS482/682 Artificial Intelligence Lecture 7: Genetic Algorithms and Constraint Satisfaction Problems 15 September 2009 Instructor: Kostas Bekris Computer.
CS 460 Spring 2011 Lecture 3 Heuristic Search / Local Search.
1 Chapter 5 Advanced Search. 2 l
Trading optimality for speed…
CSC344: AI for Games Lecture 4: Informed search
Informed search algorithms
Search CSE When you can’t use A* Hill-climbing Simulated Annealing Other strategies 2 person- games.
INTRODUÇÃO AOS SISTEMAS INTELIGENTES Prof. Dr. Celso A.A. Kaestner PPGEE-CP / UTFPR Agosto de 2011.
An Introduction to Artificial Life Lecture 4b: Informed Search and Exploration Ramin Halavati In which we see how information.
Local Search Algorithms This lecture topic Chapter Next lecture topic Chapter 5 (Please read lecture topic material before and after each lecture.
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
1 Chapter 5 Advanced Search. 2 Chapter 5 Contents l Constraint satisfaction problems l Heuristic repair l The eight queens problem l Combinatorial optimization.
Informed search algorithms
Informed search algorithms
1 Shanghai Jiao Tong University Informed Search and Exploration.
Informed search algorithms Chapter 4. Best-first search Idea: use an evaluation function f(n) for each node –estimate of "desirability"  Expand most.
Artificial Intelligence for Games Online and local search
Local Search Algorithms
Local Search Pat Riddle 2012 Semester 2 Patricia J Riddle Adapted from slides by Stuart Russell,
CHAPTER 4, Part II Oliver Schulte Summer 2011 Local Search.
Princess Nora University Artificial Intelligence Chapter (4) Informed search algorithms 1.
4/11/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 4, 4/11/2005 University of Washington, Department of Electrical Engineering Spring 2005.
A General Introduction to Artificial Intelligence.
Feng Zhiyong Tianjin University Fall  Best-first search  Greedy best-first search  A * search  Heuristics  Local search algorithms  Hill-climbing.
Best-first search Idea: use an evaluation function f(n) for each node –estimate of "desirability"  Expand most desirable unexpanded node Implementation:
Local search algorithms In many optimization problems, the state space is the space of all possible complete solutions We have an objective function that.
Local Search Algorithms This lecture topic (two lectures) Chapter Next lecture topic Chapter 5 (Please read lecture topic material before and after.
Announcement "A note taker is being recruited for this class. No extra time outside of class is required. If you take clear, well-organized notes, this.
Backtracking & Brute Force Optimization Intro2CS – weeks
Local Search. Systematic versus local search u Systematic search  Breadth-first, depth-first, IDDFS, A*, IDA*, etc  Keep one or more paths in memory.
Lecture 6 – Local Search Dr. Muhammad Adnan Hashmi 1 24 February 2016.
Local Search Algorithms and Optimization Problems
Chapter 3.5 and 3.6 Heuristic Search Continued. Review:Learning Objectives Heuristic search strategies –Best-first search –A* algorithm Heuristic functions.
CPSC 420 – Artificial Intelligence Texas A & M University Lecture 5 Lecturer: Laurie webster II, M.S.S.E., M.S.E.e., M.S.BME, Ph.D., P.E.
Department of Computer Science Lecture 5: Local Search
Local Search Algorithms CMPT 463. When: Tuesday, April 5 3:30PM Where: RLC 105 Team based: one, two or three people per team Languages: Python, C++ and.
CMPT 463. What will be covered A* search Local search Game tree Constraint satisfaction problems (CSP)
Constraints Satisfaction Edmondo Trentin, DIISM. Constraint Satisfaction Problems: Local Search In many optimization problems, the path to the goal is.
Local search algorithms In many optimization problems, the path to the goal is irrelevant; the goal state itself is the solution State space = set of "complete"
Alternative Search Formulations and Applications
Local Search Algorithms
Local Search Algorithms
Local Search Local search algorithms try to improve a given solution by modifying it   Constructive Algorithms Improvement Algorithms Need to specify:
Artificial Intelligence (CS 370D)
Back Tracking.
Analysis and design of algorithm
Informed search algorithms
Artificial Intelligence
The N-Queens Problem Search The N-Queens Problem Most slides from Milos Hauskrecht.
First Exam 18/10/2010.
Local Search Algorithms
Midterm Review.
CSC 380: Design and Analysis of Algorithms
Beyond Classical Search
Local Search Algorithms
Presentation transcript:

Local Search Algorithms Chapter 4

Outline Hill-climbing search Simulated annealing search Local beam search Genetic algorithms Ant Colony Optimization Tabu Search

Local search algorithms In many optimization problems, the path to the goal is irrelevant; the goal state itself is the solution State space = set of "complete" configurations Find configuration satisfying constraints, e.g., n- queens In such cases, we can use local search algorithms keep a single "current" state, try to improve it. Very memory efficient (only remember current state)

Example: n-queens Put n queens on an n × n board with no two queens on the same row, column, or diagonal Note that a state cannot be an incomplete configuration with m<n queens

Hill-climbing search: 8-queens problem h = number of pairs of queens that are attacking each other, either directly or indirectly (h = 17 for the above state) Each number indicates h if we move a queen in its corresponding column

Hill-climbing search: 8-queens problem A local minimum with h = 1 what can you do to get out of this local minima?)

Hill-climbing in Continuous Spaces Problem: depending on initial state, can get stuck in local maxima

Gradient Descent Assume we have some cost-function: and we want minimize over continuous variables X1,X2,..,Xn 1. Compute the gradient : 2. Take a small step downhill in the direction of the gradient: 3. Check if 4. If true then accept move, if not reject. 5. Repeat.

Exercise Describe the gradient descent algorithm for the cost function:

Line Search In GD you need to choose a step-size. Line search picks a direction, v, (say the gradient direction) and searches along that direction for the optimal step: Repeated doubling can be used to effectively search for the optimal step: There are many methods to pick search direction v. Very good method is “conjugate gradients”.

Want to find the roots of f(x). To do that, we compute the tangent at Xn and compute where it crosses the x-axis. Optimization: find roots of Does not always converge & sometimes unstable. If it converges, it converges very fast Basins of attraction for x5 − 1 = 0; darker means more iterations to converge. Newton’s Method

Simulated annealing search Idea: escape local maxima by allowing some "bad" moves but gradually decrease their frequency. This is like smoothing the cost landscape.

Properties of simulated annealing search One can prove: If T decreases slowly enough, then simulated annealing search will find a global optimum with probability approaching 1 (however, this may take VERY long) Widely used in VLSI layout, airline scheduling, etc.

Tabu Search A simple local search but with a memory. Recently visited states are added to a tabu-list and are temporarily excluded from being visited again. This way, the solver moves away from already explored regions and (in principle) avoids getting stuck in local minima.

Local beam search Keep track of k states rather than just one. Start with k randomly generated states. At each iteration, all the successors of all k states are generated. If any one is a goal state, stop; else select the k best successors from the complete list and repeat.

Genetic algorithms A successor state is generated by combining two parent states Start with k randomly generated states (population) A state is represented as a string over a finite alphabet (often a string of 0s and 1s) Evaluation function (fitness function). Higher values for better states. Produce the next generation of states by selection, crossover, and mutation

Fitness function: number of non-attacking pairs of queens (min = 0, max = 8 × 7/2 = 28) P(child) = 24/( ) = 31% P(child) = 23/( ) = 29% etc fitness: #non-attacking queens probability of being regenerated in next generation

Ant Colony Optimization Ant colony acts as a multi-agent system where the agents cooperate. Ants need to find short paths to food sources. Consider the travelling salesman problem. Construct “solutions” by joining city-to-cite steps. Inference: At each step an ant picks from its allowed set of (leftover) cities according to: Learning: pheromone forgetting: pheromone strengthening:

Linear Programming Problems of the sort: Very efficient “off-the-shelves” solvers are available for LRs. They can solve large problems with thousands of variables.