The Naïve Bayes (NB) Classifier

Slides:



Advertisements
Similar presentations
Perceptron Lecture 4.
Advertisements

Introduction to Neural Networks Computing
G53MLE | Machine Learning | Dr Guoping Qiu
NEURAL NETWORKS Perceptron
Computer Vision Lecture 18: Object Recognition II
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
5/16/2015Intelligent Systems and Soft Computing1 Introduction Introduction Hebbian learning Hebbian learning Generalised Hebbian learning algorithm Generalised.
Artificial neural networks:
Adaptive Resonance Theory (ART) networks perform completely unsupervised learning. Their competitive learning algorithm is similar to the first (unsupervised)
Neural Networks (II) Simple Learning Rule
Simple Neural Nets For Pattern Classification
A Review: Architecture
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
November 19, 2009Introduction to Cognitive Science Lecture 20: Artificial Neural Networks I 1 Artificial Neural Network (ANN) Paradigms Overview: The Backpropagation.
Un Supervised Learning & Self Organizing Maps Learning From Examples
November 9, 2010Neural Networks Lecture 16: Counterpropagation 1 Unsupervised Learning So far, we have only looked at supervised learning, in which an.
September 21, 2010Neural Networks Lecture 5: The Perceptron 1 Supervised Function Approximation In supervised learning, we train an ANN with a set of vector.
September 14, 2010Neural Networks Lecture 3: Models of Neurons and Neural Networks 1 Visual Illusions demonstrate how we perceive an “interpreted version”
November 5, 2009Introduction to Cognitive Science Lecture 16: Symbolic vs. Connectionist AI 1 Symbolism vs. Connectionism There is another major division.
An Illustrative Example
September 16, 2010Neural Networks Lecture 4: Models of Neurons and Neural Networks 1 Capabilities of Threshold Neurons By choosing appropriate weights.
November 2, 2010Neural Networks Lecture 14: Radial Basis Functions 1 Cascade Correlation Weights to each new hidden node are trained to maximize the covariance.
November 30, 2010Neural Networks Lecture 20: Interpolative Associative Memory 1 Associative Networks Associative networks are able to store a set of patterns.
September 23, 2010Neural Networks Lecture 6: Perceptron Learning 1 Refresher: Perceptron Training Algorithm Algorithm Perceptron; Start with a randomly.
Neural Networks Lecture 17: Self-Organizing Maps
November 21, 2012Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms III 1 Learning in the BPN Gradients of two-dimensional functions:
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Artificial Intelligence Lecture No. 28 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Multi-Layer Perceptrons Michael J. Watts
Mestrado em Ciência de Computadores Mestrado Integrado em Engenharia de Redes e Sistemas Informáticos VC 14/15 – TP19 Neural Networks & SVMs Miguel Tavares.
1 Machine Learning The Perceptron. 2 Heuristic Search Knowledge Based Systems (KBS) Genetic Algorithms (GAs)
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
From Biological to Artificial Neural Networks Marc Pomplun Department of Computer Science University of Massachusetts at Boston
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Artificial Intelligence & Neural Network
Mehdi Ghayoumi MSB rm 132 Ofc hr: Thur, a Machine Learning.
Introduction to Neural Networks. Biological neural activity –Each neuron has a body, an axon, and many dendrites Can be in one of the two states: firing.
Neural Network Basics Anns are analytical systems that address problems whose solutions have not been explicitly formulated Structure in which multiple.
Chapter 2 Single Layer Feedforward Networks
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
November 21, 2013Computer Vision Lecture 14: Object Recognition II 1 Statistical Pattern Recognition The formal description consists of relevant numerical.
Perceptrons Michael J. Watts
Artificial Intelligence Methods Neural Networks Lecture 3 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
1 Technological Educational Institute Of Crete Department Of Applied Informatics and Multimedia Intelligent Systems Laboratory.
1 Neural networks 2. 2 Introduction: Neural networks The nervous system contains 10^12 interconnected neurons.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
March 31, 2016Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms I 1 … let us move on to… Artificial Neural Networks.
April 5, 2016Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II 1 Capabilities of Threshold Neurons By choosing appropriate.
1 Neural Networks Winter-Spring 2014 Instructor: A. Sahebalam Instructor: A. Sahebalam Neural Networks Lecture 3: Models of Neurons and Neural Networks.
Supervised Learning in ANNs
Chapter 2 Single Layer Feedforward Networks
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Chapter 3. Artificial Neural Networks - Introduction -
Neuro-Computing Lecture 4 Radial Basis Function Network
Perceptron as one Type of Linear Discriminants
Introduction to Artificial Intelligence Lecture 11: Machine Evolution
Artificial Intelligence Lecture No. 28
Artificial Neural Networks
Capabilities of Threshold Neurons
Advanced Artificial Intelligence Classification
Prepared by: Mahmoud Rafeek Al-Farra
Neuro-Computing Lecture 2 Single-Layer Perceptrons
Artificial Intelligence 9. Perceptron
Computer Vision Lecture 19: Object Recognition III
What is Artificial Intelligence?
Presentation transcript:

The Naïve Bayes (NB) Classifier For two events A and B, Bayes’ theorem tells us the following: 𝑃(𝐴|𝐵)= 𝑃 𝐵 𝐴 𝑃 𝐵 ∙𝑃(𝐴) In other words, if an event B occurs, and we know how likely that event is to occur and how likely it occurs if event A happens at the same time, then we can update our knowledge on the probability of A occurring. We can update the prior probability P(A) to the posterior probability P(A|B) after event B took place. November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

The Naïve Bayes (NB) Classifier During training, we determine the probability distributions of all features for each of the classes. When classifying an input, we keep track of the probabilities for this input to belong to each of the classes. Typically, before analyzing the input vector, it is equally likely to belong to each class. Then we look at the value of the first feature and update the probability for each class according to Bayes’ rule and the probability distributions obtained during training. November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

The Naïve Bayes (NB) Classifier Then we do the same for the second feature, and so on, until all features have been processed. The class with the greatest final probability is then chosen as the classification result. Both training and classification can be performed very efficiently. This classifier performs well if the input features are independent (which it “naively” assumes) and their probability distributions can be accurately estimated. November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

The Artificial Neural Network (ANN) Classifier NNs are able to learn by adapting their connectivity patterns so that the organism improves its behavior in terms of reaching certain (evolutionary) goals. The strength of a connection, or whether it is excitatory or inhibitory, depends on the state of a receiving neuron’s synapses. The NN achieves learning by appropriately adapting the states of its synapses. November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

Supervised Function Approximation In supervised learning, we train an artificial NN (ANN) with a set of vector pairs, so-called exemplars. Each pair (x, y) consists of an input vector x and a corresponding output vector y. Whenever the network receives input x, we would like it to provide output y. The exemplars thus describe the function that we want to “teach” our network. Besides learning the exemplars, we would like our network to generalize, that is, give plausible output for inputs that the network had not been trained with. November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

An Artificial Neuron November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

Linear Separability Input space in the two-dimensional case (n = 2): x2 1 2 3 -3 -2 -1 x2 1 2 3 -3 -2 -1 x2 1 2 3 -3 -2 -1 1 1 1 x1 1 2 3 -3 -2 -1 x1 1 2 3 -3 -2 -1 x1 1 2 3 -3 -2 -1 w1 = 1, w2 = 2,  = 2 w1 = -2, w2 = 1,  = 2 w1 = -2, w2 = 1,  = 1 November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II

Linear Separability So by varying the weights and the threshold, we can realize any linear separation of the input space into a region that yields output 1, and another region that yields output 0. As we have seen, a two-dimensional input space can be divided by any straight line. A three-dimensional input space can be divided by any two-dimensional plane. In general, an n-dimensional input space can be divided by an (n-1)-dimensional plane or hyperplane. Of course, for n > 3 this is hard to visualize. November 8, 2018 Introduction to Artificial Intelligence Lecture 17: Neural Network Paradigms II