Introduction to Machine Learning

Slides:



Advertisements
Similar presentations
© Negnevitsky, Pearson Education, Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Advertisements

NEURAL NETWORKS Perceptron
Neural Network I Week 7 1. Team Homework Assignment #9 Read pp. 327 – 334 and the Week 7 slide. Design a neural network for XOR (Exclusive OR) Explore.
Unsupervised learning. Summary from last week We explained what local minima are, and described ways of escaping them. We investigated how the backpropagation.
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Kohonen Self Organising Maps Michael J. Watts
Artificial neural networks:
Kostas Kontogiannis E&CE
Artificial Neural Networks - Introduction -
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Machine Learning Neural Networks
Artificial Intelligence (CS 461D)
Slides are based on Negnevitsky, Pearson Education, Lecture 8 Artificial neural networks: Unsupervised learning n Introduction n Hebbian learning.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Lecture 09 Clustering-based Learning
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial neural networks:
Artificial neural networks:
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Machine Learning. Learning agent Any other agent.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
Using Neural Networks in Database Mining Tino Jimenez CS157B MW 9-10:15 February 19, 2009.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Chapter 9 Neural Network.
Artificial Neural Network Unsupervised Learning
Waqas Haider Khan Bangyal. Multi-Layer Perceptron (MLP)
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
NEURAL NETWORKS FOR DATA MINING
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
CS364 Artificial Intelligence Machine Learning Matthew Casey.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
10/17/2015Intelligent Systems and Soft Computing1 Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works Introduction,
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Introduction to Artificial Intelligence (G51IAI) Dr Rong Qu Neural Networks.
Artificial Intelligence & Neural Network
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Lecture 5 Neural Control
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Dr.Abeer Mahmoud ARTIFICIAL INTELLIGENCE (CS 461D) Dr. Abeer Mahmoud Computer science Department Princess Nora University Faculty of Computer & Information.
Each neuron has a threshold value Each neuron has weighted inputs from other neurons The input signals form a weighted sum If the activation level exceeds.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
“Principles of Soft Computing, 2 nd Edition” by S.N. Sivanandam & SN Deepa Copyright  2011 Wiley India Pvt. Ltd. All rights reserved. CHAPTER 2 ARTIFICIAL.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
 Negnevitsky, Pearson Education, Lecture 7 Artificial neural networks: Supervised learning n Introduction, or how the brain works n The neuron.
1 Artificial Intelligence CS364 Machine Learning Matthew Casey
Self-Organizing Network Model (SOM) Session 11
Machine Learning Artificial Intelligence
Learning in Neural Networks
Artificial Intelligence (CS 370D)
Artificial neural networks:
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
Machine Learning. Support Vector Machines A Support Vector Machine (SVM) can be imagined as a surface that creates a boundary between points of data.
CS364 Artificial Intelligence Machine Learning
Machine Learning. Support Vector Machines A Support Vector Machine (SVM) can be imagined as a surface that creates a boundary between points of data.
Machine Learning. Support Vector Machines A Support Vector Machine (SVM) can be imagined as a surface that creates a boundary between points of data.
Artificial Intelligence Lecture No. 28
The Network Approach: Mind as a Web
Presentation transcript:

Introduction to Machine Learning 30th October 2006 Dr Bogdan L. Vrusias b.vrusias@surrey.ac.uk

Contents Learning Artificial Neural Networks Supervised Learning Unsupervised Learning 30th October 2006 Bogdan L. Vrusias © 2006

What is Learning? ‘The action of receiving instruction or acquiring knowledge’. ‘A process which leads to the modification of behaviour or the acquisition of new abilities or responses, and which is additional to natural development by growth or maturation’. Source: Oxford English Dictionary Online: http://www.oed.com/, accessed October 2003. 30th October 2006 Bogdan L. Vrusias © 2006

Machine Learning Negnevitsky: Callan: Luger: ‘In general, machine learning involves adaptive mechanisms that enable computers to learn from experience, learn by example and learn by analogy’ (2005:165) Callan: ‘A machine or software tool would not be viewed as intelligent if it could not adapt to changes in its environment’ (2003:225) Luger: ‘Intelligent agents must be able to change through the course of their interactions with the world’ (2002:351) Learning capabilities can improve the performance of an intelligent system over time. The most popular approaches to machine learning are artificial neural networks and genetic algorithms. 30th October 2006 Bogdan L. Vrusias © 2006

Types of Learning Inductive learning Evolutionary/genetic learning Learning from examples Evolutionary/genetic learning Shaping a population of individual solutions through survival of the fittest Emergent learning Learning through social interaction: game of life 30th October 2006 Bogdan L. Vrusias © 2006

Inductive Learning Supervised learning Unsupervised learning Training examples with a known classification from a teacher Unsupervised learning No pre-classification of training examples Competitive learning: learning through competition on training examples 30th October 2006 Bogdan L. Vrusias © 2006

Key Concepts Learn from experience To adapt Generalisation Through examples, analogy or discovery To adapt Changes in response to interaction Generalisation To use experience to form a response to novel situations 30th October 2006 Bogdan L. Vrusias © 2006

Generalisation 30th October 2006 Bogdan L. Vrusias © 2006

Uses of Machine Learning Techniques and algorithms that adapt through experience. Used for: Interpretation / visualisation: summarise data Prediction: time series / stock data Classification: malignant or benign tumours Regression: curve fitting Discovery: data mining / pattern discovery 30th October 2006 Bogdan L. Vrusias © 2006

Why Machine Learning? Complexity of task / amount of data Other techniques fail or are computationally expensive Problems that cannot be defined Discovery of patterns / data mining Knowledge Engineering Bottleneck ‘Cost and difficulty of building expert systems using traditional […] techniques’ (Luger 2002:351) 30th October 2006 Bogdan L. Vrusias © 2006

Common Techniques Least squares Decision trees Support vector machines Boosting Neural networks K-means Genetic algorithms 30th October 2006 Bogdan L. Vrusias © 2006

Decision Trees A map of the reasoning process, good at solving classification problems (Negnevitsky, 2005) A decision tree represents a number of different attributes and values Nodes represent attributes Branches represent values of the attributes Path through a tree represents a decision Tree can be associated with rules 30th October 2006 Bogdan L. Vrusias © 2006

Example 1 Consider one rule for an ice-cream seller (Callan 2003:241) IF Outlook = Sunny AND Temperature = Hot THEN Sell 30th October 2006 Bogdan L. Vrusias © 2006

Example 1 Branch Root node Node Leaf Outlook Temperature Sunny Hot Sell Root node Holiday Season Overcast No Don’t Sell Node Leaf Don’t Sell Sell Yes No Mild Holiday Season Cold Don’t Sell Yes Temperature Hot Cold Mild Don’t Sell Sell 30th October 2006 Bogdan L. Vrusias © 2006

Construction Concept learning: Inducing concepts from examples We can intuitively construct a decision tree for a small set of examples Different algorithms used to construct a tree based upon the examples Most popular ID3 (Quinlan, 1986) 30th October 2006 Bogdan L. Vrusias © 2006

Which Tree? Different trees can be constructed from the same set of examples Which tree is the best? Based upon choice of attributes at each node in the tree A split in the tree (branches) should correspond to the predictor with the maximum separating power Examples can be contradictory Real-life is noisy 30th October 2006 Bogdan L. Vrusias © 2006

Extracting Rules We can extract rules from decision trees Create one rule for each root-to-leaf path Simplify by combining rules Other techniques are not so transparent: Neural networks are often described as ‘black boxes’ – it is difficult to understand what the network is doing Extraction of rules from trees can help us to understand the decision process 30th October 2006 Bogdan L. Vrusias © 2006

Issues Use prior knowledge where available Not all the examples may be needed to construct a tree Test generalisation of tree during training and stop when desired performance is achieved Prune the tree once constructed Examples may be noisy Examples may contain irrelevant attributes 30th October 2006 Bogdan L. Vrusias © 2006

Artificial Neural Networks An artificial neural network (or simply a neural network) can be defined as a model of reasoning based on the human brain. The brain consists of a densely interconnected set of nerve cells, or basic information-processing units, called neurons. The human brain incorporates nearly 10 billion neurons and 60 trillion connections, synapses, between them. By using multiple neurons simultaneously, the brain can perform its functions much faster than the fastest computers in existence today. 30th October 2006 Bogdan L. Vrusias © 2006

Artificial Neural Networks Each neuron has a very simple structure, but an army of such elements constitutes a tremendous processing power. A neuron consists of a cell body, soma, a number of fibers called dendrites, and a single long fiber called the axon. 30th October 2006 Bogdan L. Vrusias © 2006

Artificial Neural Networks Our brain can be considered as a highly complex, non-linear and parallel information-processing system. Learning is a fundamental and essential characteristic of biological neural networks. The ease with which they can learn led to attempts to emulate a biological neural network in a computer. 30th October 2006 Bogdan L. Vrusias © 2006

Artificial Neural Networks An artificial neural network consists of a number of very simple processors, also called neurons, which are analogous to the biological neurons in the brain. The neurons are connected by weighted links passing signals from one neuron to another. The output signal is transmitted through the neuron’s outgoing connection. The outgoing connection splits into a number of branches that transmit the same signal. The outgoing branches terminate at the incoming connections of other neurons in the network. 30th October 2006 Bogdan L. Vrusias © 2006

The Perceptron The operation of Rosenblatt’s perceptron is based on the McCulloch and Pitts neuron model. The model consists of a linear combiner followed by a hard limiter. The weighted sum of the inputs is applied to the hard limiter, which produces an output equal to +1 if its input is positive and 1 if it is negative. 30th October 2006 Bogdan L. Vrusias © 2006

The Perceptron How does the perceptron learn its classification tasks? This is done by making small adjustments in the weights to reduce the difference between the actual and desired outputs of the perceptron. The initial weights are randomly assigned, usually in the range [-0.5, 0.5], and then updated to obtain the output consistent with the training examples. 30th October 2006 Bogdan L. Vrusias © 2006

Multilayer neural networks A multilayer perceptron is a feedforward neural network with one or more hidden layers. The network consists of an input layer of source neurons, at least one middle or hidden layer of computational neurons, and an output layer of computational neurons. The input signals are propagated in a forward direction on a layer-by-layer basis. 30th October 2006 Bogdan L. Vrusias © 2006

What does the middle layer hide? A hidden layer “hides” its desired output. Neurons in the hidden layer cannot be observed through the input/output behaviour of the network. There is no obvious way to know what the desired output of the hidden layer should be. Commercial ANNs incorporate three and sometimes four layers, including one or two hidden layers. Each layer can contain from 10 to 1000 neurons. Experimental neural networks may have five or even six layers, including three or four hidden layers, and utilise millions of neurons. 30th October 2006 Bogdan L. Vrusias © 2006

Supervised Learning Supervised or active learning is learning with an external “teacher” or a supervisor who presents a training set to the network. The most populat supervised neural network is the back-propagation neural network. 30th October 2006 Bogdan L. Vrusias © 2006

Back-propagation neural network Learning in a multilayer network proceeds the same way as for a perceptron. A training set of input patterns is presented to the network. The network computes its output pattern, and if there is an error – or in other words a difference between actual and desired output patterns – the weights are adjusted to reduce this error. 30th October 2006 Bogdan L. Vrusias © 2006

Back-propagation neural network 30th October 2006 Bogdan L. Vrusias © 2006

Back-propagation neural network Network represented by McCulloch-Pitts model for solving the Exclusive-OR operation. 30th October 2006 Bogdan L. Vrusias © 2006

Back-propagation neural network (a) Decision boundary constructed by hidden neuron 3; (b) Decision boundary constructed by hidden neuron 4; (c) Decision boundaries constructed by the complete three-layer network 30th October 2006 Bogdan L. Vrusias © 2006

Unsupervised Learning Unsupervised or self-organised learning does not require an external teacher. During the training session, the neural network receives a number of different input patterns, discovers significant features in these patterns and learns how to classify input data into appropriate categories. Unsupervised learning tends to follow the neuro-biological organisation of the brain. Most popular unsupervised neural networks are the Hebbian network and the Self-Organising Feature Map. 30th October 2006 Bogdan L. Vrusias © 2006

Hebbian Network Hebb’s Law can be represented in the form of two rules: If two neurons on either side of a connection are activated synchronously, then the weight of that connection is increased. If two neurons on either side of a connection are activated asynchronously, then the weight of that connection is decreased. 30th October 2006 Bogdan L. Vrusias © 2006

Competitive Learning In competitive learning, neurons compete among themselves to be activated. While in Hebbian learning, several output neurons can be activated simultaneously, in competitive learning, only a single output neuron is active at any time. The output neuron that wins the “competition” is called the winner-takes-all neuron. Self-organising feature maps are based on competitive learning. 30th October 2006 Bogdan L. Vrusias © 2006

What is a self-organising feature map? Our brain is dominated by the cerebral cortex, a very complex structure of billions of neurons and hundreds of billions of synapses. The cortex includes areas that are responsible for different human activities (motor, visual, auditory, somatosensory, etc.), and associated with different sensory inputs. We can say that each sensory input is mapped into a corresponding area of the cerebral cortex. The cortex is a self-organising computational map in the human brain. The self-organising feature map has been introduced by Teuvo Kohonen and therefore is also called Kohonen network. 30th October 2006 Bogdan L. Vrusias © 2006

What is a self-organising feature map? 30th October 2006 Bogdan L. Vrusias © 2006

The Kohonen network The Kohonen model provides a topological mapping. It places a fixed number of input patterns from the input layer into a higher-dimensional output or Kohonen layer. Training in the Kohonen network begins with the winner’s neighbourhood of a fairly large size. Then, as training proceeds, the neighbourhood size gradually decreases. 30th October 2006 Bogdan L. Vrusias © 2006

The Kohonen network The lateral connections are used to create a competition between neurons. The neuron with the largest activation level among all neurons in the output layer becomes the winner. This neuron is the only neuron that produces an output signal. The activity of all other neurons is suppressed in the competition. The lateral feedback connections produce excitatory or inhibitory effects, depending on the distance from the winning neuron. 30th October 2006 Bogdan L. Vrusias © 2006

Competitive learning in the Kohonen network To illustrate competitive learning, consider the Kohonen network with 100 neurons arranged in the form of a two-dimensional lattice with 10 rows and 10 columns. The network is required to classify two-dimensional input vectors each neuron in the network should respond only to the input vectors occurring in its region. The network is trained with 1000 two-dimensional input vectors generated randomly in a square region in the interval between –1 and +1. 30th October 2006 Bogdan L. Vrusias © 2006

Competitive learning in the Kohonen network 30th October 2006 Bogdan L. Vrusias © 2006