Summary So Far Extremes in classes of games: –Nonadversarial, perfect information, deterministic –Adversarial, imperfect information, chance  Adversarial,

Slides:



Advertisements
Similar presentations
Adversarial Search Chapter 6 Sections 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Advertisements

Adversarial Search Chapter 6 Section 1 – 4. Types of Games.
Adversarial Search We have experience in search where we assume that we are the only intelligent being and we have explicit control over the “world”. Lets.
Decision Making Under Uncertainty CSE 495 Resources: –Russell and Norwick’s book.
Games & Adversarial Search Chapter 5. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent’s reply. Time.
February 7, 2006AI: Chapter 6: Adversarial Search1 Artificial Intelligence Chapter 6: Adversarial Search Michael Scherger Department of Computer Science.
CMSC 671 Fall 2001 Class #8 – Thursday, September 27.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2008.
Adversarial Search Chapter 6 Section 1 – 4.
Adversarial Search Chapter 5.
COMP-4640: Intelligent & Interactive Systems Game Playing A game can be formally defined as a search problem with: -An initial state -a set of operators.
1 Game Playing. 2 Outline Perfect Play Resource Limits Alpha-Beta pruning Games of Chance.
Adversarial Search: Game Playing Reading: Chapter next time.
Lecture 12 Last time: CSPs, backtracking, forward checking Today: Game Playing.
Adversarial Search CSE 473 University of Washington.
Adversarial Search Chapter 6.
Adversarial Search 對抗搜尋. Outline  Optimal decisions  α-β pruning  Imperfect, real-time decisions.
An Introduction to Artificial Intelligence Lecture VI: Adversarial Search (Games) Ramin Halavati In which we examine problems.
1 Adversarial Search Chapter 6 Section 1 – 4 The Master vs Machine: A Video.
10/19/2004TCSS435A Isabelle Bichindaritz1 Game and Tree Searching.
Game Playing 최호연 이춘우. Overview Intro: Games as search problems Perfect decisions in 2-person games Imperfect decisions Alpha-beta pruning.
G51IAI Introduction to AI Minmax and Alpha Beta Pruning Garry Kasparov and Deep Blue. © 1997, GM Gabriel Schwartzman's Chess Camera, courtesy IBM.
CMSC 463 Chapter 5: Game Playing Prof. Adam Anthony.
Minimax and Alpha-Beta Reduction Borrows from Spring 2006 CS 440 Lecture Slides.
Lecture 13 Last time: Games, minimax, alpha-beta Today: Finish off games, summary.
Games with Chance 2012/04/25 1. Nondeterministic Games: Backgammon White moves clockwise toward 25. Black moves counterclockwise toward 0. A piece can.
This time: Outline Game playing The minimax algorithm
CS 561, Sessions Last time: search strategies Uninformed: Use only information available in the problem formulation Breadth-first Uniform-cost Depth-first.
Game Playing CSC361 AI CSC361: Game Playing.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering CSCE 580 Artificial Intelligence Ch.6: Adversarial Search Fall 2008 Marco Valtorta.
1 DCP 1172 Introduction to Artificial Intelligence Lecture notes for Chap. 6 [AIMA] Chang-Sheng Chen.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2006.
Adversarial Search: Game Playing Reading: Chess paper.
Games & Adversarial Search Chapter 6 Section 1 – 4.
Game Playing: Adversarial Search Chapter 6. Why study games Fun Clear criteria for success Interesting, hard problems which require minimal “initial structure”
Game Playing State-of-the-Art  Checkers: Chinook ended 40-year-reign of human world champion Marion Tinsley in Used an endgame database defining.
CSC 412: AI Adversarial Search
PSU CS 370 – Introduction to Artificial Intelligence Game MinMax Alpha-Beta.
Lecture 6: Game Playing Heshaam Faili University of Tehran Two-player games Minmax search algorithm Alpha-Beta pruning Games with chance.
Complexity and Emergence in Games (Ch. 14 & 15). Seven Schemas Schema: Conceptual framework concentrating on one aspect of game design Schemas: –Games.
1 Game Playing Why do AI researchers study game playing? 1.It’s a good reasoning problem, formal and nontrivial. 2.Direct comparison with humans and other.
Games as Game Theory Systems (Ch. 19). Game Theory It is not a theoretical approach to games Game theory is the mathematical study of decision making.
Game-playing AIs Part 1 CIS 391 Fall CSE Intro to AI 2 Games: Outline of Unit Part I (this set of slides)  Motivation  Game Trees  Evaluation.
Adversarial Search Chapter 6 Section 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Notes on Game Playing by Yun Peng of theYun Peng University of Maryland Baltimore County.
Introduction to Artificial Intelligence CS 438 Spring 2008 Today –AIMA, Ch. 6 –Adversarial Search Thursday –AIMA, Ch. 6 –More Adversarial Search The “Luke.
Axioms Let W be statements known to be true in a domain An axiom is a rule presumed to be true An axiomatic set is a collection of axioms Given an axiomatic.
1 Adversarial Search CS 171/271 (Chapter 6) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Adversarial Search Chapter 6 Section 1 – 4. Search in an Adversarial Environment Iterative deepening and A* useful for single-agent search problems What.
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
Paula Matuszek, CSC 8520, Fall Based in part on aima.eecs.berkeley.edu/slides-ppt 1 CS 8520: Artificial Intelligence Adversarial Search Paula Matuszek.
ARTIFICIAL INTELLIGENCE (CS 461D) Princess Nora University Faculty of Computer & Information Systems.
Adversarial Search Chapter 6 Section 1 – 4. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time.
Explorations in Artificial Intelligence Prof. Carla P. Gomes Module 5 Adversarial Search (Thanks Meinolf Sellman!)
Adversarial Search Chapter 5 Sections 1 – 4. AI & Expert Systems© Dr. Khalid Kaabneh, AAU Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
ADVERSARIAL SEARCH Chapter 6 Section 1 – 4. OUTLINE Optimal decisions α-β pruning Imperfect, real-time decisions.
Chapter 5 Adversarial Search. 5.1 Games Why Study Game Playing? Games allow us to experiment with easier versions of real-world situations Hostile agents.
Uncertainty and Games (Ch. 15). Uncertainty If a game outcome is certain can it achieve meaningful play? –Example of such a game? Two kinds of uncertainty:
5/4/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 9, 5/4/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Last time: search strategies
Adversarial Search Chapter 5.
Expectimax Lirong Xia. Expectimax Lirong Xia Project 2 MAX player: Pacman Question 1-3: Multiple MIN players: ghosts Extend classical minimax search.
Adversarial Search.
Expectimax Lirong Xia.
Game Playing Fifth Lecture 2019/4/11.
Adversarial Search Chapter 6 Section 1 – 4.
Presentation transcript:

Summary So Far Extremes in classes of games: –Nonadversarial, perfect information, deterministic –Adversarial, imperfect information, chance  Adversarial, perfect information, deterministic Adversarial, perfect information, deterministic  Minimax trees  Optimal playing strategy if tree is finite  But requires generating whole tree  But there are workarounds  Cut-off and evaluation functions  α-β pruning

Chance (Non-game) Applications: Negotiation Auctions Military planning

Example: Blackgammon Game overview: –Goal is to move all one’s pieces off the board. –Requires all pieces to be in the home board. –White moves clockwise toward 25, and black counterclockwise toward 0. –A piece can move everywhere except if there are several opponent pieces there. –Roll dice at the beginning of a player’s turn to determine legal moves.

Example Configuration legal moves: 5-10, , , , 11-16

Detour: Probability Suppose that I flip a “fair” coin:  what is the probability that it will come heads:  Expected gain if you bet $X on heads: 0.5 Suppose that I flip a “totally unfair” coin (always come heads):  what is the probability that it will come heads:  Expected gain if you bet $X on heads: 1 $X Maximum Expected Utility principle MEU([p 1,S 1 ; p 2,S 2 ; … ; p n,S n ]) =  i p i U(S i ) $X/2

Example Suppose that you are in a TV show and you have already earned 1’ so far. Now, the presentator propose you a gamble: he will flip a coin if the coin comes up heads you will earn 3’ But if it comes up tails you will loose the 1’ What do you decide? First shot: U(winning $X) = X MEU ([0.5,0; 0.5,3’ ]) = 1’ This utility is called the expected monetary value

Example (II) If we use the expected monetary value of the lottery does it take the bet? Yes!, because: MEU([0.5,0; 0.5,3’ ]) = 1’ > MEU([1,1’ ; 0,3’ ]) = 1’ But is this really what you would do? Not me!

Example (III) Second shot: Let S = “my current wealth” S’ = “my current wealth” + $1’ S’’ = “my current wealth” + $3’ MEU(Accept) = MEU(Decline) = 0.5U(S) + 0.5U(S’’) U(S’) 0.5U(S) + 0.5U(S’’) U(S’) If U(S) = 5, U(S’) = 8, U(S’’) = 10, would you accept the bet? No! = 7.5 = 8 $ U

Human Judgment and Utility Decision theory is a normative theory: describe how agents should act Experimental evidence suggest that people violate the axioms of utility Tversky and Kahnerman (1982) and Allen (1953):  Experiment with people  Choice was given between A and B and then between C and D: A: 80% chance of $4000 B: 100% chance of $3000 C: 20% chance of $4000 D: 25% chance of $3000

Human Judgment and Utility (II) Majority choose B over A and C over D If U($0) = 0 MEU([0.8,4000; 0.2,0]) = MEU([1,3000; 0,4000]) = 0.8U($4000) U($3000) Thus, 0.8U($4000) < U($3000) MEU([0.2,4000; 0.8,0]) = MEU([0.25,3000; 0.65, 0]) = 0.2U($4000) 0.25U($3000) Thus, 0.2U($4000) > 0.25U($3000) Thus, there cannot be no utility function consistent with these values

Human Judgment and Utility (III) The point is that it is very hard to model an automatic agent that behaves like a human (back to the Turing test) However, the utility theory does give some formal way of model decisions and as such is used to generate consistent decisions

Extending Minimax Trees: Expectiminimax Chance node denote possible dice rolls Each branch from a chance node is labeled with the probability that the branch will be taken  If distribution is uniform then probability is 1/n, where n is the number of choices Each position has an expected utility

Expected Utility: Expectimax If node n is terminal, EY(n) = utility(n) If n is a nonterminal node: expectimax and expectimin Terminal MAX Dice C Expectimax(C) =  i p(d i )max S  S(C,di) (utility(s)) where S(C,d i ) is the set of all legal moves for P(d i ) P(1,1) P(6,6)

Expected Utility: MIN, MAX Terminal MAX Dice C MIN(M) = min C  children(M) (Expectimax(C)) (that is, apply standard minimax-value formula) MIN M

Expected Utility: Expectimin Terminal MAX Dice C MIN M Dice C’ Expectimin(C’) =  i p(d i )min S  S(C,di) (utility(s)) where S(C,d i ) is the set of all legal moves for P(d i )

Closing Notes These trees can be very large, therefore Cut-off and evaluation functions –Evaluation functions have to be linear functions: EF(state) = w 1 f 1 (state) + w 2 f 2 (state) + … + w n f n (state) Complexity –Minimax (i.e., w/o chance): O(b m ) –Expectiminimax: O(b m n m ), where n is the number of distinct roles