Computer Go : A Go player Rohit Gurjar CS365 Project Proposal, IIT Kanpur Guided By – Prof. Amitabha Mukerjee.

Slides:



Advertisements
Similar presentations
Adversarial Search Chapter 6 Sections 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Advertisements

Learning in Computer Go David Silver. The Problem Large state space  Approximately states  Game tree of about nodes  Branching factor.
Adversarial Search Chapter 6 Section 1 – 4. Types of Games.
Adversarial Search We have experience in search where we assume that we are the only intelligent being and we have explicit control over the “world”. Lets.
Games & Adversarial Search Chapter 5. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent’s reply. Time.
February 7, 2006AI: Chapter 6: Adversarial Search1 Artificial Intelligence Chapter 6: Adversarial Search Michael Scherger Department of Computer Science.
Prepare for 4x4x4 tic-tac-toe
Adversarial Search Chapter 5.
Chapter 6: The physical symbol system hypothesis
An Introduction to Artificial Intelligence Lecture VI: Adversarial Search (Games) Ramin Halavati In which we examine problems.
1 Adversarial Search Chapter 6 Section 1 – 4 The Master vs Machine: A Video.
10/19/2004TCSS435A Isabelle Bichindaritz1 Game and Tree Searching.
Games CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Minimax and Alpha-Beta Reduction Borrows from Spring 2006 CS 440 Lecture Slides.
Problem Solving Using Search Reduce a problem to one of searching a graph. View problem solving as a process of moving through a sequence of problem states.
This time: Outline Game playing The minimax algorithm
The Move Decision Strategy of Indigo Author: Bruno Bouzy Presented by: Ling Zhao University of Alberta March 7, 2007.
Honte, a Go-Playing Program Using Neural Nets Frederik Dahl.
Maximizing the Chance of Winning in Searching Go Game Trees Presenter: Ling Zhao March 16, 2005 Author: Keh-Hsun Chen Accepted by Information Sciences.
Solving Probabilistic Combinatorial Games Ling Zhao & Martin Mueller University of Alberta September 7, 2005 Paper link:
Learning Shape in Computer Go David Silver. A brief introduction to Go Black and white take turns to place down stones Once played, a stone cannot move.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering CSCE 580 Artificial Intelligence Ch.6: Adversarial Search Fall 2008 Marco Valtorta.
Reinforcement Learning of Local Shape in the Game of Atari-Go David Silver.
Evolution and Coevolution of Artificial Neural Networks playing Go Thesis by Peter Maier, Salzburg, April 2004 Additional paper used Computer Go, by Martin.
1 Solving Ponnuki-Go on Small Board Paper: Solving Ponnuki-Go on small board Authors: Erik van der Werf, Jos Uiterwijk, Jaap van den Herik Presented by:
Adversarial Search: Game Playing Reading: Chess paper.
Monte Carlo Go Has a Way to Go Haruhiro Yoshimoto (*1) Kazuki Yoshizoe (*1) Tomoyuki Kaneko (*1) Akihiro Kishimoto (*2) Kenjiro Taura (*1) (*1)University.
Othello Sean Farrell June 29, Othello Two-player game played on 8x8 board All pieces have one white side and one black side Initial board setup.
THE RENJU GAME BY ABHISHEK JAIN, PRANSHU GUPTA & RHYTHM DAS PCLUB SUMMER PROJECT PRESENTATION JUNE, L7 IIT KANPUR MENTOR – SANIL JAIN.
Reinforcement Learning of Local Shape in the Game of Atari-Go David Silver.
Neural Networks Slides by Megan Vasta. Neural Networks Biological approach to AI Developed in 1943 Comprised of one or more layers of neurons Several.
Go An ancient Oriental board game Andrew Simons. Introduction 2 player game of skill. Popular in the Far East, growing in the West. Simple rules, extremely.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
C. Benatti, 3/15/2012, Slide 1 GA/ICA Workshop Carla Benatti 3/15/2012.
Introduction Many decision making problems in real life
Othello Artificial Intelligence With Machine Learning
Multi-Layer Perceptrons Michael J. Watts
Development of a Machine-Learning-Based AI For Go By Justin Park.
 Summary  How to Play Go  Project Details  Demo  Results  Conclusions.
Temperature Discovery Search Temperature Discovery Search (TDS) is a new minimaxbased game tree search method designed to compute or approximate the temperature.
1 CO Games Development 1 Week 11 Search Methods Gareth Bellaby.
Adversarial Search Chapter 6 Section 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Computer Go : A Go player Rohit Gurjar CS365 Project Presentation, IIT Kanpur Guided By – Prof. Amitabha Mukerjee.
CSC Intro. to Computing Lecture 22: Artificial Intelligence.
Amazons Experiments in Computer Amazons, Martin Mueller and Theodore Tegos, 2002 Exhaustive Search in the Game Amazons Raymond Georg Snatzke, 2002 Presented.
CHECKERS: TD(Λ) LEARNING APPLIED FOR DETERMINISTIC GAME Presented By: Presented To: Amna Khan Mis Saleha Raza.
Blondie24 Presented by Adam Duffy and Josh Hill. Overview Introduction to new concepts Design of Blondie24 Testing and results Other approaches to checkers.
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
Feature selection with Neural Networks Dmitrij Lagutin, T Variable Selection for Regression
Solving Kriegspiel endings with brute force: the case of KR vs. K Paolo Ciancarini Gian Piero Favini University of Bologna.
Parallelization in Computer Board Games Ian Princivalli.
AI & Machine Learning Libraries By Logan Kearsley.
Learning to Play the Game of GO Lei Li Computer Science Department May 3, 2007.
Artificial Intelligence
Adversarial Search Chapter 6 Section 1 – 4. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time.
A Scalable Machine Learning Approach to Go Pierre Baldi and Lin Wu UC Irvine.
Explorations in Artificial Intelligence Prof. Carla P. Gomes Module 5 Adversarial Search (Thanks Meinolf Sellman!)
Dimensions of Neural Networks Ali Akbar Darabi Ghassem Mirroshandel Hootan Nokhost.
ConvNets for Image Classification
Adversarial Search Chapter 5 Sections 1 – 4. AI & Expert Systems© Dr. Khalid Kaabneh, AAU Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
ADVERSARIAL SEARCH Chapter 6 Section 1 – 4. OUTLINE Optimal decisions α-β pruning Imperfect, real-time decisions.
Understanding AI of 2 Player Games. Motivation Not much experience in AI (first AI project) and no specific interests/passion that I wanted to explore.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Understanding AlphaGo. Go Overview Originated in ancient China 2,500 years ago Two players game Goal - surround more territory than the opponent 19X19.
Artificial Intelligence AIMA §5: Adversarial Search
Stochastic tree search and stochastic games
AlphaGo with Deep RL Alpha GO.
Training a Neural Network
CIS 488/588 Bruce R. Maxim UM-Dearborn
Adversarial Search Chapter 6 Section 1 – 4.
Presentation transcript:

Computer Go : A Go player Rohit Gurjar CS365 Project Proposal, IIT Kanpur Guided By – Prof. Amitabha Mukerjee

Introduction to the game A two player board game Around 3000 years old Players have white and black stones A stone (or a group) is captured if it is completely surrounded by enemy stones Goal is to surround territories (empty points or enemy stones) by with one’s own stones Image ref :

Past work done A lot of work has been done to develop a Computer Go player But the best Go programs are ranked only as 1-3 kyu (equivalent to a intermediate amateur) Go is an extremely difficult game for computers to play well. Four main factor causes failure of standard AI techniques 1) The state space of the game of go is intractably huge. 2) The branching factor of go is intractably huge. 3) It is difficult to construct a heuristic function to evaluate board states. 4) Proper go strategy requires an extremely deep look-ahead. Ref : A machine learning approach to computer Go (2007) – Jeffrey Bagdis

Algorithms used for Move Generation Two components were necessary – Evaluation Function – A tree search Finding a good evaluation function is very hard Solution- Generate goals Choose a specific EF related to one goal and associate a specific move generator with this goal. (Reduced complexity) What if more than one goal is relevant to winning the game? Computer Go, an AI oriented survey- Bruno Bouzy, Tristan Cazenave (2000)

Tree search In classical games tree search uses the EF to find out the best possible move High branching factor is a problem (~200) Local TS – Selecting moves localized on a part of the board – Problems Defining locality criterion Given the results of local TS, reconstruction of the global result Independence of local situations

Tree search Splitting the game into sub-games was proved to be useful in the endgame. * Problem – finding independent sub-games Alpha-Beta pruning. – as the tree being searched, cut off some sections of the tree from consideration, without at all reducing the optimality of accuracy of an exhaustive search. * M.Müller, Decomposition search: A combinatorial games approach to game tree search, with applications to solving Go endgames (1999)

Machine Learning Previous algorithms are inadequate for developing a Go player. Machine learning through neural network technique A function(nonlinear) maps a vector(board state) to a scalar(0-1) which shows the probability of winning The perceptron - simplified computational model of the biological neurons Ref : A machine learning approach to computer Go (2007) – Jeffrey Bagdis

Perceptron A single perceptron can be trained to produce a certain output for a corresponding input by adjusting its input weights appropriately. In order to achieve a more precise approximation, perceptrons can be connected together to form Multi-layer networks - perceptrons can be connected together to form multi-layer networks Ref : A machine learning approach to computer Go (2007) – Jeffrey Bagdis

Multilayer Network Ref : A machine learning approach to computer Go (2007) – Jeffrey Bagdis

Conclusion This approach seems effective at producing an approximation of a value function that can be used to play go. But this value function is not by itself sufficient for strong play Combination of this with some of the previously shown algorithms may give good results.

Algorithms used for go Finding a good evaluation function is very hard – Concrete EF With the high branching factor of Go computers cannot search down to the depth of terminating the game – Conceptual EF used by program Indigo To observe human players, and capture the useful concepts