Game Playing: Adversarial Search Chapter 6. Why study games Fun Clear criteria for success Interesting, hard problems which require minimal “initial structure”

Slides:



Advertisements
Similar presentations
CMSC 471 Spring 2014 Class #9 Tue 2/25/14 Game Playing Professor Marie desJardins,
Advertisements

Game Playing CS 63 Chapter 6
Adversarial Search Chapter 6 Section 1 – 4. Types of Games.
Adversarial Search We have experience in search where we assume that we are the only intelligent being and we have explicit control over the “world”. Lets.
February 7, 2006AI: Chapter 6: Adversarial Search1 Artificial Intelligence Chapter 6: Adversarial Search Michael Scherger Department of Computer Science.
Artificial Intelligence Adversarial search Fall 2008 professor: Luigi Ceccaroni.
CMSC 671 Fall 2001 Class #8 – Thursday, September 27.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2008.
CS 484 – Artificial Intelligence
Adversarial Search Chapter 6 Section 1 – 4.
University College Cork (Ireland) Department of Civil and Environmental Engineering Course: Engineering Artificial Intelligence Dr. Radu Marinescu Lecture.
Adversarial Search Chapter 5.
Adversarial Search: Game Playing Reading: Chapter next time.
Lecture 12 Last time: CSPs, backtracking, forward checking Today: Game Playing.
Games CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
CMSC 463 Chapter 5: Game Playing Prof. Adam Anthony.
This time: Outline Game playing The minimax algorithm
Lecture 02 – Part C Game Playing: Adversarial Search
1 Game Playing Chapter 6 Additional references for the slides: Luger’s AI book (2005). Robert Wilensky’s CS188 slides:
Game Playing CSC361 AI CSC361: Game Playing.
1 search CS 331/531 Dr M M Awais A* Examples:. 2 search CS 331/531 Dr M M Awais 8-Puzzle f(N) = g(N) + h(N)
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2006.
Adversarial Search: Game Playing Reading: Chess paper.
Games & Adversarial Search Chapter 6 Section 1 – 4.
ICS-270a:Notes 5: 1 Notes 5: Game-Playing ICS 270a Winter 2003.
Game Playing State-of-the-Art  Checkers: Chinook ended 40-year-reign of human world champion Marion Tinsley in Used an endgame database defining.
1 Adversary Search Ref: Chapter 5. 2 Games & A.I. Easy to measure success Easy to represent states Small number of operators Comparison against humans.
CSC 412: AI Adversarial Search
CHAPTER 6 : ADVERSARIAL SEARCH
PSU CS 370 – Introduction to Artificial Intelligence Game MinMax Alpha-Beta.
Game Playing Chapter 5. Game playing §Search applied to a problem against an adversary l some actions are not under the control of the problem-solver.
Lecture 6: Game Playing Heshaam Faili University of Tehran Two-player games Minmax search algorithm Alpha-Beta pruning Games with chance.
Game Playing.
AD FOR GAMES Lecture 4. M INIMAX AND A LPHA -B ETA R EDUCTION Borrows from Spring 2006 CS 440 Lecture Slides.
Game Playing Chapter 5. Game playing §Search applied to a problem against an adversary l some actions are not under the control of the problem-solver.
1 Computer Group Engineering Department University of Science and Culture S. H. Davarpanah
Chapter 12 Adversarial Search. (c) 2000, 2001 SNU CSE Biointelligence Lab2 Two-Agent Games (1) Idealized Setting  The actions of the agents are interleaved.
Chapter 6 Adversarial Search. Adversarial Search Problem Initial State Initial State Successor Function Successor Function Terminal Test Terminal Test.
Adversarial Search CS311 David Kauchak Spring 2013 Some material borrowed from : Sara Owsley Sood and others.
Game-playing AIs Part 1 CIS 391 Fall CSE Intro to AI 2 Games: Outline of Unit Part I (this set of slides)  Motivation  Game Trees  Evaluation.
Notes on Game Playing by Yun Peng of theYun Peng University of Maryland Baltimore County.
Games. Adversaries Consider the process of reasoning when an adversary is trying to defeat our efforts In game playing situations one searches down the.
Adversarial Search Chapter 6 Section 1 – 4. Search in an Adversarial Environment Iterative deepening and A* useful for single-agent search problems What.
Game Playing. Introduction One of the earliest areas in artificial intelligence is game playing. Two-person zero-sum game. Games for which the state space.
Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri CS 440 / ECE 448 Introduction to Artificial Intelligence.
Paula Matuszek, CSC 8520, Fall Based in part on aima.eecs.berkeley.edu/slides-ppt 1 CS 8520: Artificial Intelligence Adversarial Search Paula Matuszek.
ARTIFICIAL INTELLIGENCE (CS 461D) Princess Nora University Faculty of Computer & Information Systems.
Game-playing AIs Part 2 CIS 391 Fall CSE Intro to AI 2 Games: Outline of Unit Part II  The Minimax Rule  Alpha-Beta Pruning  Game-playing.
CMSC 421: Intro to Artificial Intelligence October 6, 2003 Lecture 7: Games Professor: Bonnie J. Dorr TA: Nate Waisbrot.
Game Playing: Adversarial Search chapter 5. Game Playing: Adversarial Search  Introduction  So far, in problem solving, single agent search  The machine.
Game Playing: Adversarial Search Chapter 6. Why study games Fun Clear criteria for success Offer an opportunity to study problems involving {hostile,
Adversarial Search 2 (Game Playing)
Adversarial Search and Game Playing Russell and Norvig: Chapter 6 Slides adapted from: robotics.stanford.edu/~latombe/cs121/2004/home.htm Prof: Dekang.
Explorations in Artificial Intelligence Prof. Carla P. Gomes Module 5 Adversarial Search (Thanks Meinolf Sellman!)
Adversarial Search 1 (Game Playing)
Adversarial Search In this lecture, we introduce a new search scenario: game playing 1.two players, 2.zero-sum game, (win-lose, lose-win, draw) 3.perfect.
Adversarial Search and Game Playing. Topics Game playing Game trees Minimax Alpha-beta pruning Examples.
Chapter 5 Adversarial Search. 5.1 Games Why Study Game Playing? Games allow us to experiment with easier versions of real-world situations Hostile agents.
Game Playing Why do AI researchers study game playing?
Adversarial Search and Game-Playing
PENGANTAR INTELIJENSIA BUATAN (64A614)
Adversarial Search and Game Playing (Where making good decisions requires respecting your opponent) R&N: Chap. 6.
Game Playing in AI by: Gaurav Phapale 05 IT 6010
CMSC 471 Fall 2011 Class #8-9 Tue 9/27/11 – Thu 9/29/11 Game Playing
Game Playing: Adversarial Search
CMSC 671 Fall 2010 Tue 9/21/10 Game Playing
Game Playing: Adversarial Search
Game Playing Chapter 5.
Presentation transcript:

Game Playing: Adversarial Search Chapter 6

Why study games Fun Clear criteria for success Interesting, hard problems which require minimal “initial structure” Games often define very large search spaces –chess 10*120 nodes Historical reasons Offer an opportunity to study problems involving {hostile, adversarial, competing} agents. Different from games studied in game theory

Typical Case (perfect games) 2-person game Players alternate moves Zero-sum-- one players loss is the other’s gain. Perfect information -- both players have access to complete information about the state of the game. No information is hidden from either player. No chance (e.g., using dice) involved Clear rules for legal moves (n o uncertain position transition involved) Well-defined outcomes (W/L/D) Examples: Tic-Tac-Toe, Checkers, Chess, Go, Nim, Othello Not: Bridge, Solitaire, Backgammon,...

How to play a game A way to play such a game is to: –Consider all the legal moves you can make. –Each move leads to a new board configuration (position). –Evaluate each resulting position and determine which is best –Make that move. –Wait for your opponent to move and repeat? Key problems are: –Representing the “board” –Generating all legal next boards –Evaluating a position –Look ahead

Game Trees Problem spaces for typical games represented as trees. Root node represents the “board” configuration at which a decision must be made as to what is the best single move to make next. (not necessarily the initial configuration) Evaluator function rates a board position. f(board) (a real number). Arcs represent the possible legal moves for a player (no costs associates to arcs Terminal nodes represent end-game configurations (the result must be one of “win”, “lose”, and “draw”, possibly with numerical payoff)

If it is my turn to move, then the root is labeled a "MAX" node; otherwise it is labeled a "MIN" node indicating my opponent's turn. Each level of the tree has nodes that are all MAX or all MIN; nodes at level i are of the opposite kind from those at level i+1 Complete game tree: includes all configurations that can be generated from the root by legal moves (all leaves are terminal nodes) Incomplete game tree: includes all configurations that can be generated from the root by legal moves to a given depth (looking ahead to a given steps)

Evaluation Function Evaluation function or static evaluator –Evaluates the "goodness" of a game position. –Contrast with heuristic search where the evaluation function was a non-negative estimate of the cost from the start node to a goal and passing through the given node. The zero-sum assumption allows us to use a single evaluation function to describe the goodness of a board with respect to both players. –f(n) > 0: position n good for me and bad for you. –f(n) < 0: position n bad for me and good for you –f(n) near 0: position n is a neutral position. –f(n) >> 0: win for me. –f(n) << 0: win for you..

Evaluation function is a heuristic function, and it is where the domain experts’ knowledge resides. Example of an Evaluation Function for Tic-Tac-Toe: f(n) = [# of 3-lengths open for me] - [# of 3-lengths open for you] where a 3-length is a complete row, column, or diagonal. Alan Turing’s function for chess –f(n) = w(n)/b(n) where w(n) = sum of the point value of white’s pieces and b(n) is sum for black. Most evaluation functions are specified as a weighted sum of position features: f(n) = w1*feat1(n) + w2*feat2(n) wn*featk(n) Example features for chess are piece count, piece placement, squares controlled, etc. Deep Blue has about 6,000 features in its evaluation function.

An example (partial) game tree for Tic-Tac-Toe - f(n) = +1 if the position is a win for X. f(n) = -1 if the position is a win for O. f(n) = 0 if the position is a draw.

Some Chess Positions and their Evaluations

Minimax Rule Goal of game tree search: to determine one move for Max player that maximizes the guaranteed payoff for a given game tree for MAX Regardless of the moves the MIN will take The value of each node (Max and MIN) is determined by (back up from) the values of its children MAX plays the worst case scenario: Always assume MIN to take moves to maximize his pay-off (i.e., to minimize the pay-off of MAX) For a MAX node, the backed up value is the maximum of the values associated with its children For a MIN node, the backed up value is the minimum of the values associated with its children

Minimax Tree MAX node MIN node f valueA 1 is selected as the next move

Minimax procedure Create start node as a MAX node with current board configuration Expand nodes down to some depth (i.e., ply) of lookahead in the game. Apply the evaluation function at each of the leaf nodes Obtain the “back up" values for each of the non-leaf nodes from its children by Minimax rule until a value is computed for the root node. Pick the operator associated with the child node whose backed up value determined the value at the root as the move for MAX

Minimax Search MAX MIN This is the move selected by minimax Static evaluator value

Comments on Minimax search The search is depth-first with the given depth (ply) as the limit –Time complexity: O(b^d) –Linear space complexity Performance depends on –Quality of evaluation functions (domain knowledge) –Depth of the search (computer power and search algorithm) Different from ordinary state space search –Not to search for a complete solution but for one move only –No cost is associated with each arc –MAX does not know how MIN is going to counter each of his moves Minimax rule is a basis for other game tree search algorithms

Alpha-beta pruning We can improve on the performance of the minimax algorithm through alpha-beta pruning. Basic idea: “If you have an idea that is surely bad, don't take the time to see how truly awful it is.” -- Pat Winston 271 =2 >=2 <=1 ? We don’t need to compute the value at this node. No matter what it is it can’t effect the value of the root node.

Alpha-beta pruning Traverse the search tree in depth-first order At each Max node n, alpha(n) = maximum value found so far –Start with -infinity and only increase –Increases if a child of n returns a value greater than the current alpha –Serve as a tentative lower bound of the final pay-off At each Min node n, beta(n) = minimum value found so far –Start with infinity and only decrease –Decreases if a child of n returns a value less than the current beta –Serve as a tentative upper bound of the final pay-off

Alpha-beta pruning Alpha cutoff: Given a Max node n, cutoff the search below n (i.e., don't generate or examine any more of n's children) if alpha(n) >= beta(n) (alpha increases and passes beta from below) Beta cutoff.: Given a Min node n, cutoff the search below n (i.e., don't generate or examine any more of n's children) if beta(n) <= alpha(n) (beta decreases and passes alpha from above) Carry alpha and beta values down during search Pruning occurs whenever alpha >= beta

Alpha-beta search

Alpha-beta algorithm Two functions recursively call each other function MAX-value (n, alpha, beta) if n is a leaf node then return f(n); for each child n’ of n do alpha :=max{alpha, MIN-value(n’, alpha, beta)}; if alpha >= beta then return beta /* pruning */ end{do} return alpha function MIN-value (n, alpha, beta) if n is a leaf node then return f(n); for each child n’ of n do beta :=min{beta, MAX-value(n’, alpha, beta)}; if beta <= alpha then return alpha /* pruning */ end{do} return beta

Effectiveness of Alpha-beta pruning Alpha-Beta is guaranteed to compute the same value for the root node as computed by Minimax. Worst case: NO pruning, examining O(b^d) leaf nodes, where each node has b children and a d-ply search is performed Best case: examine only O(b^(d/2)) leaf nodes. –You can search twice as deep as Minimax! Or the branch factor is b^(1/2) rather than b. Best case is when each player's best move is the leftmost alternative, i.e. at MAX nodes the child with the largest value generated first, and at MIN nodes the child with the smallest value generated first. In Deep Blue, they found empirically that with Alpha-Beta pruning the average branching factor at each node was about 6 instead of about 35-40

An example of Alpha-beta pruning max min

Final tree max min