Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Multi-Layer Perceptron (MLP)
A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
NEURAL NETWORKS Perceptron
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Machine Learning Neural Networks
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
How does the mind process all the information it receives?
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
An Introduction To The Backpropagation Algorithm Who gets the credit?
CS 4700: Foundations of Artificial Intelligence
November 21, 2012Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms III 1 Learning in the BPN Gradients of two-dimensional functions:
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Dr. Hala Moushir Ebied Faculty of Computers & Information Sciences
Modelling Language Evolution Lecture 2: Learning Syntax Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks
Explorations in Neural Networks Tianhui Cai Period 3.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition LECTURE 16: NEURAL NETWORKS Objectives: Feedforward.
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
CSC321: Neural Networks Lecture 2: Learning with linear neurons Geoffrey Hinton.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Multi-Layer Perceptron
Neural Networks Steven Le. Overview Introduction Architectures Learning Techniques Advantages Applications.
Akram Bitar and Larry Manevitz Department of Computer Science
EE459 Neural Networks Backpropagation
Introduction to Neural Networks. Biological neural activity –Each neuron has a body, an axon, and many dendrites Can be in one of the two states: firing.
CS621 : Artificial Intelligence
Neural Networks - Berrin Yanıkoğlu1 Applications and Examples From Mitchell Chp. 4.
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Neural Networks 2nd Edition Simon Haykin
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
An Introduction To The Backpropagation Algorithm.
Artificial Neural Networks This is lecture 15 of the module `Biologically Inspired Computing’ An introduction to Artificial Neural Networks.
Learning: Neural Networks Artificial Intelligence CMSC February 3, 2005.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Supervised Learning in ANNs
Learning with Perceptrons and Neural Networks
CS623: Introduction to Computing with Neural Nets (lecture-5)
CSE 473 Introduction to Artificial Intelligence Neural Networks
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Prof. Carolina Ruiz Department of Computer Science
CSC 578 Neural Networks and Deep Learning
network of simple neuron-like computing elements
Capabilities of Threshold Neurons
Lecture Notes for Chapter 4 Artificial Neural Networks
CS623: Introduction to Computing with Neural Nets (lecture-5)
Computer Vision Lecture 19: Object Recognition III
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
Akram Bitar and Larry Manevitz Department of Computer Science
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit

Course Overview Learning Introduction to neural nets Learning syntax Evolution Syntax Learning bias and structure Culture Iterated learning The Talking Heads (practical)

Computers for modelling Computers in linguistics Engineering (speech and language technologies) Research tools (waveform analysis, psycholinguistic stimuli etc.) Recently: modelling building Why build models? Why use computers? What is a model anyway?

What is a model? One view: We use models when we can’t be sure what our theories predict Especially useful when dealing with complex systems THEORY MODEL PREDICTION OBSERVATION

A simple example Vowels exist in a “space” Only some patterns arise cross-linguistically E.g. vowel space seems to be symmetrically filled Why?

Theory to Model We need a theory to explain vowel-space universal Possible theory: Vowels tend to avoid being close to each other to maintain perceptual distinctiveness. Use model to test theory (Liljencrants & Lindblom 1972) In general, computational models are useful when dealing with “complex systems”

Is language a complex system? Cultural evolution Individual learning Biological evolution Yes – evolution on many different timescales: Computational models will help us understand these interactions…

Learning Language learning is crucial to language evolution What is learning? Learning occurs when an organism changes its internal state on the basis of experience What do we need to model learning? 1.a model of internal states 2.A model of experience 3.An algorithm to change 1 into 2

One approach: Neural nets An approach to internal states based on the brain An artificial neuron is a computational unit that sums inputs and uses them to decide whether to produce an output

Networks of neurons Typically there will be many connected neurons Information is stored in weights on the connections Weights multiply signals sent between nodes Signals into a node can be excitatory or inhibitory

An artificial neuron Add up all the inputs multiplied by their weights f(net) is the “activation function” that scales the input

A useful activation function All or nothing for big excitations or inhibitions… … but more sensitive in between.

AND: a very simple network A network that works out if both inputs are activated: INPUT 1INPUT 2 BIAS NODE (always set to 1.0) OUTPUT Network gives an output over 0.5 only if both inputs are 1.

OR: another very simple network A network that works out if either input is activated: INPUT 1INPUT 2 BIAS NODE (always set to 1.0) OUTPUT Network gives an output over 0.5 if either input is 1.

XOR: a difficult challenge A network that works out if only one input is activated: INPUT 1INPUT 2 BIAS NODE (always set to 1.0) OUTPUT ?? ? Solution needs more complex net with three layers. WHY?

XOR network - step 1 XOR is the same as OR but not AND Calculate OR Calculate NOT AND AND the results NOT ANDOR AND

XOR network - step 2 OUTPUT BIAS NODE HIDDEN 1HIDDEN 2 INPUT 1INPUT NOT ANDOR AND

But what about learning? We now have: a model of internal states (connection weights) a model of experience (inputs and outputs) Learning: set the weights in response to experience How? Compare network behaviour with “correct” behaviour Adjust the weights to reduce network error

Error-driven learning 1.Set weights to random values 2.Present input pattern 3.Feed-forward activation through the network to get an output 4.Calculate difference between output and desired output (i.e. error) 5.Adjust weights so that the error is reduced 6.Repeat until network is producing the desired results.

Gradient descent Gradient descent is a form of error-driven learning Start on random point of “error surface” Move on surface in direction of steepest slope Potential problems: May overshoot the global minimum Might get stuck in local minimum

Example: learning past tense of verbs Network that takes present tense form of verb… …and produces past tense. Uses examples to set weights Generalises to add /-ed/ to verbs it’s never seen before. Has it learnt a linguistic rule?

Is this psychologically plausible? We need an error signal Where does this error signal come from? Possibilities: A teacher Reinforcement The outcome of some prediction: e.g. what’s the next word? what’s the past tense of this verb?

Summary Modelling tests theories Computer modelling appropriate for complex systems Language evolution involves several complex systems Neural nets are one approach to modelling learning Networks can be made to adapt to data through error-driven learning Next lecture: how to model acquisition of syntax