The Principle of Presence: A Heuristic for Growing Knowledge Structured Neural Networks Laurent Orseau, INSA/IRISA, Rennes, France.

Slides:



Advertisements
Similar presentations
KULIAH II JST: BASIC CONCEPTS
Advertisements

1 Short Term Memories and Forcing the Re-use of Knowledge for Generalization Laurent Orseau INSA/IRISA Rennes, France.
Learning in Neural and Belief Networks - Feed Forward Neural Network 2001 년 3 월 28 일 안순길.
also known as the “Perceptron”
Neural Networks  A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Tuomas Sandholm Carnegie Mellon University Computer Science Department
B.Macukow 1 Neural Networks Lecture 4. B.Macukow 2 McCulloch symbolism The symbolism introduced by McCulloch at the basis of simplified Venn diagrams.
ECE 667 Synthesis & Verification - Boolean Functions 1 ECE 667 Spring 2013 ECE 667 Spring 2013 Synthesis and Verification of Digital Circuits Boolean Functions.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Neural Networks Basic concepts ArchitectureOperation.
Self-Organizing Hierarchical Neural Network
Boolean Functions and their Representations
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #30 4/15/02 Neural Networks.
AN INTERACTIVE TOOL FOR THE STOCK MARKET RESEARCH USING RECURSIVE NEURAL NETWORKS Master Thesis Michal Trna
MAE 552 Heuristic Optimization Instructor: John Eddy Lecture #31 4/17/02 Neural Networks.
MACHINE LEARNING 12. Multilayer Perceptrons. Neural Networks Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Artificial Neural Networks
Biologically Inspired Robotics Group,EPFL Associative memory using coupled non-linear oscillators Semester project Final Presentation Vlad TRIFA.
Information Fusion Yu Cai. Research Article “Comparative Analysis of Some Neural Network Architectures for Data Fusion”, Authors: Juan Cires, PA Romo,
Artificial Neural Networks -Application- Peter Andras
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
1 st Neural Network: AND function Threshold(Y) = 2 X1 Y X Y.
Stochastic Optimization and Simulated Annealing Psychology /719 January 25, 2001.
© N. Kasabov Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering, MIT Press, 1996 INFO331 Machine learning. Neural networks. Supervised.
IE 585 Introduction to Neural Networks. 2 Modeling Continuum Unarticulated Wisdom Articulated Qualitative Models Theoretic (First Principles) Models Empirical.
Multi-Layer Perceptrons Michael J. Watts
Neural Networks Ellen Walker Hiram College. Connectionist Architectures Characterized by (Rich & Knight) –Large number of very simple neuron-like processing.
Machine Learning Chapter 4. Artificial Neural Networks
Lecture 7 Topics –Boolean Algebra 1. Logic and Bits Operation Computers represent information by bit A bit has two possible values, namely zero and one.
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
CSC321: Introduction to Neural Networks and machine Learning Lecture 16: Hopfield nets and simulated annealing Geoffrey Hinton.
CHAPTER 5 S TOCHASTIC G RADIENT F ORM OF S TOCHASTIC A PROXIMATION Organization of chapter in ISSO –Stochastic gradient Core algorithm Basic principles.
1 CMSC 671 Fall 2001 Class #25-26 – Tuesday, November 27 / Thursday, November 29.
Neural Networks Steven Le. Overview Introduction Architectures Learning Techniques Advantages Applications.
Rapid integration of new schema- consistent information in the Complementary Learning Systems Theory Jay McClelland, Stanford University.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
MAKING PRODUCT DECISIONS Economics, March  Remember: we are the supplier, making decisions about what to PRODUCE!
Mutation Operator Evolution for EA-Based Neural Networks By Ryan Meuth.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Chapter 18 Connectionist Models
Warm Up Solve each inequality. 1. x + 3 ≤ x ≤ 7 23 < –2x + 3
Perceptrons Michael J. Watts
Chapter 6 Neural Network.
Properties A property is something that is true for all situations.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
CS206Evolutionary Robotics Artificial Neural Networks Input Layer Output Layer Input Layer Output Layer ++ ++
Machine Learning Supervised Learning Classification and Regression
1-4 Solving Inequalities
CSC321 Lecture 18: Hopfield nets and simulated annealing
Datastructure.
Dr. Kenneth Stanley September 6, 2006
Neural Networks A neural network is a network of simulated neurons that can be used to recognize instances of patterns. NNs learn by searching through.
Artificial Intelligence Methods
Training a Neural Network
XOR problem Input 2 Input 1
N-Gram Model Formulas Word sequences Chain rule of probability
The use of Neural Networks to schedule flow-shop with dynamic job arrival ‘A Multi-Neural Network Learning for lot Sizing and Sequencing on a Flow-Shop’
The Naïve Bayes (NB) Classifier
The Network Approach: Mind as a Web
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
Design of Experiments CHM 585 Chapter 15.
Sanguthevar Rajasekaran University of Connecticut
Properties of Numbers Review Problems.
Presentation transcript:

The Principle of Presence: A Heuristic for Growing Knowledge Structured Neural Networks Laurent Orseau, INSA/IRISA, Rennes, France

Neural Networks Efficient at learning single problems Fully connected Convergence in W 3 Lifelong learning: Specific cases can be important More knowledge, more weights Catastrophic forgetting -> Full connectivity not suitable -> Need localilty

How can people learn so fast? Focus, attention Raw table storing? Frog and Car and Running woman With generalization

What do people memorize? (1) 1 memory: a set of « things » Things are made of other, simpler things Thing=concept Basic concept=perceptual event

What do people memorize? (2) Remember only what is present in mind at the time of memorization: What is seen What is heard What is thought Etc.

What do people memorize? (3) Not what is not in mind! Too many concepts are known What is present: Few things Probably important What is absent: Many things Probably unrelevant Good but not always true -> heuristic

Presence in everyday life Easy to see what is present, harder to tell what is missing Infants lose attention to balls that have just disappeared The zero number invented long after other digits Etc.

The principle of presence Memorization = create a new concept upon only active concepts Independant of the number of known concepts Few active concepts -> few variables -> fast generalization

Implications A concept can be active or inactive. Activity must reflect importance, be rare ~ event (programming) New concept = conjunction of actives ones Concepts must be re-usable(lifelong): Re-use = create a link from this concept 2 independant concepts = 2 units -> More symbolic than MLP: a neuron can represent too many things

Implementation: NN Nonlinearity Graphs properties: local or global connectivity Weights: Smooth on-line generalization Resistant to noise But more symbolic: Inactivity: piecewise continuous activation function Knowledge not too much distributed Concepts not too much overlapping

First implementation Inputs: basic events Output: target concept No macro-concept: -> 3-layer Neuron = conjunction, unless explicit (supervised learning), -> DNF Output weights simulate priority

Locality in learning Only one neuron modified at a time: Nearest = most activated If target concept not activated when it should: Generalize the nearest connected neuron Add a neuron for that specific case If target active, but not enough or too much: Generalize the most activating neuron

Learning: example (0) Must learn AB. Examples: ABC, ABD, ABE, but not AB. A B C D E AB … Target already exists Inputs:

Learning: example (1) ABC: A B C D E AB 1 Disjunction 1/3 2/3 Conjunction /Ns N1 N1 active when A, B and C all active

Learning : example (2) ABD: A B C D E AB N1 >1/3 <1/3 1 2/3 N2 1/3 1 2/3 1/3

Learning : example (3) ABE: N1 slightly active for AB A B C D E AB N1 >>1/3 <<1/3 1 N2 1/3 1 2/3 >1/3 <1/3

Learning : example (4) Final: N1 has generalized, active for AB A B C D E AB N1 1/2 0 1 N2 1/3 1 2/3 Unuseful neuron Deleted by criterion

NETtalk task TDNN: 120 neurons, cnx, 90% Presence: 753 neurons, cnx, 74% Then learns by heart If inputs activity reversed -> catastrophic! Many cognitive tasks heavily biased toward the principle of presence?

Advantages w/r NNs As many inputs as wanted, only active ones are used Lifelong learning: Large scale networks Learns specific cases and generalizes, both quickly Can lower weights without wrong prediction -> imitation

But… Few data, limiting the number of neurons: not as good as backprop Creates many neurons (but can be deleted) No negative weights

Work in progress Negative case, must stay rare Inhibitory links Re-use of concepts Macro-concepts: each concept can become an input