Multilayer feed-forward artificial neural networks for Class-modeling F. Marini, A. Magrì, R. Bucci Dept. of Chemistry - University of Rome “La Sapienza”

Slides:



Advertisements
Similar presentations
Machine Learning: Connectionist McCulloch-Pitts Neuron Perceptrons Multilayer Networks Support Vector Machines Feedback Networks Hopfield Networks.
Advertisements

Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Machine Learning Neural Networks
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
Machine Learning Motivation for machine learning How to set up a problem How to design a learner Introduce one class of learners (ANN) –Perceptrons –Feed-forward.
Aula 4 Radial Basis Function Networks
Artificial Neural Network
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
On the Application of Artificial Intelligence Techniques to the Quality Improvement of Industrial Processes P. Georgilakis N. Hatziargyriou Schneider ElectricNational.
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Radial Basis Function Networks
Artificial Neural Networks
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Classification Part 3: Artificial Neural Networks
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Explorations in Neural Networks Tianhui Cai Period 3.
Chapter 9 Neural Network.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
ANNs (Artificial Neural Networks). THE PERCEPTRON.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 23 Nov 2, 2005 Nanjing University of Science & Technology.
Chapter 3 Neural Network Xiu-jun GONG (Ph. D) School of Computer Science and Technology, Tianjin University
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
Appendix B: An Example of Back-propagation algorithm
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
NEURAL NETWORKS FOR DATA MINING
Classification / Regression Neural Networks 2
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Artificial Neural Networks An Introduction. What is a Neural Network? A human Brain A porpoise brain The brain in a living creature A computer program.
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 31: Feedforward N/W; sigmoid.
Multi-Layer Perceptron
Non-Bayes classifiers. Linear discriminants, neural networks.
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
CS621 : Artificial Intelligence
CS344: Introduction to Artificial Intelligence (associated lab: CS386) Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 32: sigmoid neuron; Feedforward.
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
EEE502 Pattern Recognition
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Kim HS Introduction considering that the amount of MRI data to analyze in present-day clinical trials is often on the order of hundreds or.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A Document-Level Sentiment Analysis Approach Using Artificial Neural Network and Sentiment Lexicons Yan Zhu.
Evolutionary Computation Evolving Neural Network Topologies.
Multiple-Layer Networks and Backpropagation Algorithms
Artificial Neural Networks
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Structure learning with deep autoencoders
CS621: Artificial Intelligence
Prof. Carolina Ruiz Department of Computer Science
ECE 471/571 - Lecture 17 Back Propagation.
Classification Neural Networks 1
Artificial Neural Network & Backpropagation Algorithm
Synaptic DynamicsII : Supervised Learning
of the Artificial Neural Networks.
network of simple neuron-like computing elements
Capabilities of Threshold Neurons
CS621: Artificial Intelligence Lecture 22-23: Sigmoid neuron, Backpropagation (Lecture 20 and 21 taken by Anup on Graphical Models) Pushpak Bhattacharyya.
CS621: Artificial Intelligence Lecture 18: Feedforward network contd
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Multilayer feed-forward artificial neural networks for Class-modeling F. Marini, A. Magrì, R. Bucci Dept. of Chemistry - University of Rome “La Sapienza”

The starting question…. Despite literature on NNs has increased significantly, no paper considers the possibility of performing class- modeling

class modeling: what…. Class modeling considers one class at a time Any object can then belong or not to that specific class model As a consequence, any object can be assigned to only one class, to more than one class or to no class at all classificationclass modeling

…..and why Flexibility Additional information: –sensitivity: fraction of samples from category X accepted by the model of category X –specificity: fraction of samples from category Y (or Z, W….) refused by the model of category X No need to rebuild the existing models each time a new category is added. less equivocal answer to the question: “are the analytical data compatible with the product being X as declared?”

A first step forward A particular kind of NN, after suitable modifications could be used for performing class-modeling ( Anal. Chim. Acta, 544 (2005), 306 ) –Kohonen SOM –Addition of dummy random vectors to the training set –Computation of a suitable (non-parametric) probability distribution after mapping on the 2D Kohonen layer. –Definition of the category space based on this distribution

In this communication… …The possibility of using a different type of neural network (multilayer feed-forward) to operate class- modeling is studied –How to? –Examples

Just a few words about NN   Sophocles

NN: a mathematical approach From a computational point of view, ANNs represent a way to operate a non-linear functional mapping between an input and an output space. This functional relation is expressed in an implicit way (via a combination of suitably weighted non-linear functions, in the case of MLF-NN ) ANNs are usually represented as groups of elementary computational units (neurons) performing simultaneously the same operations. Types of NN differ in how neurons are grouped and how they operate

Multilayer feed-forward NN Individual processing units are organized in three types of layer: input, hidden and output All neurons within the same layer operate simultaneously output hidden input x 1 x 2 x 3 x 4 x 5 y 1 y 2 y 3 y 4

The artificial neuron f(  )  w 1k w 2k w 3k x3x3 x2x2 x1x1 zkzk hidden input x 1 x 2 x 3 x 4 x 5

The artificial neuron f(  )  w 1j w 2j w 3j z3z3 z2z2 z1z1 yjyj hidden input x 1 x 2 x 3 x 4 x 5 output y 1 y 2 y 3 y 4

Training Iterative variation of connection weights, to minimize an error criterion. Usually, backpropagation algorithm is used:

MLF class-modeling: what to do? Model for each category has to be built using only training samples from that category Suitable definition of category space

Somewhere to start from When targets are equal to input values, hidden nodes could be thought of as a sort of non-linear principal components Input Hidden Input x 1 x 2 x 3 X j x m Output value of hidden node 1 Output value of hidden node 2

… and a first ending point For each category a neural network model is computed providing the input vector also as desired target vector N inp -N hid -N inp Number of hidden layer is estimated by loo-cv (minimum reconstruction error in prediction) The optimized model is then used to predict unknown samples: –Sample is presented to the network –Vector of predicted responses (which is an estimate of the original input vector) is computed –Prediction error is calculated and compared to the average prediction error for samples belonging to the category (as in SIMCA ).

NN-CM in practice Separate category autoscaling if is lower than a predifined threshold, the sample is refused by the category model.

A couple of examples

The classical X-OR 200 training samples:200 training samples: –100 class 1 –100 class test samples:200 test samples: –100 class 1 –100 class 2 3 hidden neurons for each category

Results Sensitivity: –100% class 1, 100% class2 Specificity: –75% class1 vs class 2 –67% class2 vs class 1 Prediction ability: –87% class1 –83% class2 –85% overall These results are significantly better than with SIMCA and UNEQ ( specificities lower than 30% and classification slightly higher than 60%)

A very small data-set: honey

CM of honey samples 76 samples of honey from 6 different botanical origins (honeydew, wildflower, sulla, heather, eucalyptus and chestnut) samples per class 2 input variables: specific rotation; total acidity Despite the small number of samples, a good NN model was obtained (2 hidden neurons for each class) Possibility of drawing a Coomans’ plot

Further work and Conclusions A novel approach to class-modeling based on multilayer feed-forward NN was presented Preliminary results seem to indicate its usefulness in cases where traditional class modeling fails Effect of training set dimension should be further invetigated (our “small” data set was too good to be used for obtaining a definitive answer) We are analyzing other “exotic” data sets for classification where traditional methods fail.

Acknowledgements Prof. Jure Zupan, SloveniaProf. Jure Zupan, Slovenia