CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 1 Combining Inductive and Analytical Learning Why combine inductive and analytical learning?

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Explanation-Based Learning (borrowed from mooney et al)
Analytical Learning.
Combining Inductive and Analytical Learning
Slides from: Doug Gray, David Poole
 2002, G.Tecuci, Learning Agents Laboratory 1 Learning Agents Laboratory Computer Science Department George Mason University Prof. Gheorghe Tecuci Deductive.
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Data Mining Classification: Alternative Techniques
Knowledge Representation and Reasoning Learning Sets of Rules and Analytical Learning Harris Georgiou – 4.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
Artificial Intelligence Chapter 14. Resolution in the Propositional Calculus Artificial Intelligence Chapter 14. Resolution in the Propositional Calculus.
처음 페이지로 이동 Chapter 11: Analytical Learning Inductive learning training examples n Analytical learning prior knowledge + deductive reasoning n Explanation.
Università di Milano-Bicocca Laurea Magistrale in Informatica
Machine Learning Neural Networks
1 Part I Artificial Neural Networks Sofia Nikitaki.
Artificial Neural Networks ML Paul Scheible.
Relational Data Mining in Finance Haonan Zhang CFWin /04/2003.
Hybrid Systems Two examples of the Combination of Rule-Based Systems and neural nets By Pieter Buzing.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Neural Networks Marco Loog.
1 Chapter 18 Learning from Observations Decision tree examples Additional source used in preparing the slides: Jean-Claude Latombe’s CS121 slides: robotics.stanford.edu/~latombe/cs121.
Machine Learning: Symbol-Based
Artificial Neural Networks
CSCI 5582 Fall 2006 CSCI 5582 Artificial Intelligence Lecture 22 Jim Martin.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Inductive Logic Programming Includes slides by Luis Tari CS7741L16ILP.
Machine Learning Chapter 3. Decision Tree Learning
Artificial Neural Networks
1 Machine Learning: Lecture 11 Analytical Learning / Explanation-Based Learning (Based on Chapter 11 of Mitchell, T., Machine Learning, 1997)
Machine Learning Chapter 11. Analytical Learning
Multiple-Layer Networks and Backpropagation Algorithms
1 Machine Learning What is learning?. 2 Machine Learning What is learning? “That is what learning is. You suddenly understand something you've understood.
Machine Learning Chapter 11.
Neural Networks Chapter 6 Joost N. Kok Universiteit Leiden.
Chapter 9 Neural Network.
Theory Revision Chris Murphy. The Problem Sometimes we: – Have theories for existing data that do not match new data – Do not want to repeat learning.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Friday, February 4, 2000 Lijun.
November 10, Machine Learning: Lecture 9 Rule Learning / Inductive Logic Programming.
Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture.
Machine Learning Chapter 2. Concept Learning and The General-to-specific Ordering Tom M. Mitchell.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence Monday, January 22, 2001 William.
1 CMSC 671 Fall 2001 Class #25-26 – Tuesday, November 27 / Thursday, November 29.
Machine Learning Chapter 5. Artificial IntelligenceChapter 52 Learning 1. Rote learning rote( โรท ) n. วิถีทาง, ทางเดิน, วิธีการตามปกติ, (by rote จากความทรงจำ.
Overview Concept Learning Representation Inductive Learning Hypothesis
For Monday Finish chapter 19 No homework. Program 4 Any questions?
CS 5751 Machine Learning Chapter 3 Decision Tree Learning1 Decision Trees Decision tree representation ID3 learning algorithm Entropy, Information gain.
For Monday Finish chapter 19 Take-home exam due. Program 4 Any questions?
CS Inductive Bias1 Inductive Bias: How to generalize on novel data.
CS 5751 Machine Learning Chapter 10 Learning Sets of Rules1 Learning Sets of Rules Sequential covering algorithms FOIL Induction as the inverse of deduction.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Machine Learning Concept Learning General-to Specific Ordering
Ensemble Methods in Machine Learning
CpSc 810: Machine Learning Analytical learning. 2 Copy Right Notice Most slides in this presentation are adopted from slides of text book and various.
Decision Tree Learning Presented by Ping Zhang Nov. 26th, 2007.
Data Mining and Decision Support
Introduction Machine Learning: Chapter 1. Contents Types of learning Applications of machine learning Disciplines related with machine learning Well-posed.
More Symbolic Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Chap. 10 Learning Sets of Rules 박성배 서울대학교 컴퓨터공학과.
CS623: Introduction to Computing with Neural Nets (lecture-18) Pushpak Bhattacharyya Computer Science and Engineering Department IIT Bombay.
Machine Learning Supervised Learning Classification and Regression
CS 9633 Machine Learning Explanation Based Learning
Chapter 2 Concept Learning
CS 9633 Machine Learning Concept Learning
CS 9633 Machine Learning Inductive-Analytical Methods
Resolution in the Propositional Calculus
Data Mining Lecture 11.
Biointelligence Lab School of Computer Sci. & Eng.
Machine Learning Chapter 2
Machine Learning Chapter 2
Presentation transcript:

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 1 Combining Inductive and Analytical Learning Why combine inductive and analytical learning? KBANN: prior knowledge to initialize the hypothesis TangentProp, EBNN: prior knowledge alters search objective FOCL: prior knowledge alters search operators

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 2 Inductive and Analytical Learning Inductive learning Hypothesis fits data Statistical inference Requires little prior knowledge Syntactic inductive bias Analytical learning Hypothesis fits domain theory Deductive inference Learns from scarce data Bias is domain theory

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 3 What We Would Like General purpose learning method: No domain theory  learn as well as inductive methods Perfect domain theory  learn as well as P ROLOG -EBG Accommodate arbitrary and unknown errors in domain theory Accommodate arbitrary and unknown errors in training data Inductive learning Plentiful data No prior knowledge Scarce data Perfect prior knowledge Analytical learning

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 4 Domain Theory Cup  Stable, Liftable, OpenVessel Stable  BottomIsFlat Liftable  Graspable, Light Graspable  HasHandle OpenVessel  HasConcavity, ConcavityPointsUp Cup StableLiftableOpenVessel Graspable BottomIsFlatLightHasConcavityConcavityPointsUpHasHandle

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 5 Training Examples Cups Non-Cups BottomIsFlat         ConcavityPointsUp        Expensive     Fragile       HandleOnTop   HandleOnSide    HasConcavity          HasHandle      Light         MadeOfCeramic     MadeOfPaper   MadeOfStyroForm    

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 6 KBANN Knowledge Based Artificial Neural Networks KBANN (data D, domain theory B) 1. Create a feedforward network h equivalent to B 2. Use B ACKPROP to tune h to fit D

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 7 Neural Net Equivalent to Domain Theory Expensive BottomIsFlat MadeOfCeramic MadeOfStyrofoam MadeOfPaper HasHandle HandleOnTop HandleOnSide Light HasConcavity ConcavityPointsUp Fragile Stable Liftable OpenVessel CupGraspable large positive weight large negative weight negligible weight

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 8 Creating Network Equivalent to Domain Theory Create one unit per horn clause rule (an AND unit) Connect unit inputs to corresponding clause antecedents For each non-negated antecedent, corresponding input weight w  W, where W is some constant For each negated antecedent, weight w  -W Threshold weight w 0  -(n -.5) W, where n is number of non-negated antecedents Finally, add additional connections with near-zero weights Liftable  Graspable, ¬Heavy

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 9 Result of Refining the Network Expensive BottomIsFlat MadeOfCeramic MadeOfStyrofoam MadeOfPaper HasHandle HandleOnTop HandleOnSide Light HasConcavity ConcavityPointsUp Fragile Stable Liftable OpenVessel CupGraspable large positive weight large negative weight negligible weight

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 10 KBANN Results Classifying promoter regions in DNA (leave one out testing): Backpropagation: error rate 8/106 KBANN: 4/106 Similar improvements on other classification, control tasks.

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 11 Hypothesis Space Search in KBANN Hypotheses that fit training data equally well Initial hypothesis for KBANN Initial hypothesis for Backpropagation Hypothesis Space

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 12 EBNN Explanation Based Neural Network Key idea: Previously learned approximate domain theory Domain theory represented by collection of neural networks Learn target function as another neural network

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 13 Expensive BottomIsFlat MadeOfCeramic MadeOfStyrofoam MadeOfPaper HasHandle HandleOnTop HandleOnSide Light HasConcavity ConcavityPointsUp Fragile Stable Explanation in Terms of Domain Theory Prior learned networks for useful concepts combined into a single target network Graspable Stable OpenVessel LiftableCup

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 14 TangetProp

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 15

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 16 Hypothesis Space Search in TangentProp Hypotheses that maximize fit to data TangetProp Search Backpropagation Search Hypothesis Space Hypotheses that maximize fit to data and prior knowledge

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 17 FOCL Adaptation of FOIL that uses domain theory When adding a literal to a rule, not only consider adding single terms, but also think about adding terms from domain theory May also prune specializations generated

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 18 Search in FOCL Cup  HasHandle [2+,3-] Cup  Cup  ¬HasHandle [2+,3-] Cup  Fragile [2+,4-] Cup  BottomIsFlat, Light, HasConcavity, ConcavityPointsUp [4+,2-]... Cup  BottomIsFlat, Light, HasConcavity, ConcavityPointsUp, HandleOnTop [0+,2-] Cup  BottomIsFlat, Light, HasConcavity, ConcavityPointsUp, ¬ HandleOnTop [4+,0-] Cup  BottomIsFlat, Light, HasConcavity, ConcavityPointsUp, HandleOnSide [2+,0-]...

CS 5751 Machine Learning Chapter 12 Comb. Inductive/Analytical 19 FOCL Results Recognizing legal chess endgame positions: 30 positive, 30 negative examples FOIL: 86% FOCL: 94% (using domain theory with 76% accuracy) NYNEX telephone network diagnosis 500 training examples FOIL: 90% FOCL: 98% (using domain theory with 95% accuracy)