1 Convolutional neural networks Abin - Roozgard. 2  Introduction  Drawbacks of previous neural networks  Convolutional neural networks  LeNet 5 

Slides:



Advertisements
Similar presentations
Face Recognition: A Convolutional Neural Network Approach
Advertisements

Convolutional Neural Networks
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Combining Inductive and Analytical Learning Ch 12. in Machine Learning Tom M. Mitchell 고려대학교 자연어처리 연구실 한 경 수
Reading for Next Week Textbook, Section 9, pp A User’s Guide to Support Vector Machines (linked from course website)
Thesis title: “Studies in Pattern Classification – Biological Modeling, Uncertainty Reasoning, and Statistical Learning” 3 parts: (1)Handwritten Digit.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Neural NetworksNN 11 Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Chapter 5 NEURAL NETWORKS
Rutgers CS440, Fall 2003 Neural networks Reading: Ch. 20, Sec. 5, AIMA 2 nd Ed.
Multi Layer Perceptrons (MLP) Course website: The back-propagation algorithm Following Hertz chapter 6.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Introduction to Recurrent neural networks (RNN), Long short-term memory (LSTM) Wenjie Pei In this coffee talk, I would like to present you some basic.
CHAPTER 11 Back-Propagation Ming-Feng Yeh.
NEURAL NETWORKS Introduction
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Pattern Recognition Vidya Manian Dept. of Electrical and Computer Engineering University of Puerto Rico INEL 5046, Spring 2007
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
MSE 2400 EaLiCaRA Spring 2015 Dr. Tom Way
Multi Layer NN and Bit-True Modeling of These Networks SILab presentation Ali Ahmadi September 2007.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
Explorations in Neural Networks Tianhui Cai Period 3.
ANNs (Artificial Neural Networks). THE PERCEPTRON.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Matlab Matlab Sigmoid Sigmoid Perceptron Perceptron Linear Linear Training Training Small, Round Blue-Cell Tumor Classification Example Small, Round Blue-Cell.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
COMPARISON OF IMAGE ANALYSIS FOR THAI HANDWRITTEN CHARACTER RECOGNITION Olarik Surinta, chatklaw Jareanpon Department of Management Information System.
1 Chapter 11 Neural Networks. 2 Chapter 11 Contents (1) l Biological Neurons l Artificial Neurons l Perceptrons l Multilayer Neural Networks l Backpropagation.
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
EE459 Neural Networks Examples of using Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University.
Analysis of Classification Algorithms In Handwritten Digit Recognition Logan Helms Jon Daniele.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Mentor Prof. Amitabha Mukerjee Deepak Pathak Kaustubh Tapi 10346
Each neuron has a threshold value Each neuron has weighted inputs from other neurons The input signals form a weighted sum If the activation level exceeds.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Machine Learning Artificial Neural Networks MPλ ∀ Stergiou Theodoros 1.
Optical Character Recognition
Artificial Neural Networks By: Steve Kidos. Outline Artificial Neural Networks: An Introduction Frank Rosenblatt’s Perceptron Multi-layer Perceptron Dot.
CSE343/543 Machine Learning Mayank Vatsa Lecture slides are prepared using several teaching resources and no authorship is claimed for any slides.
Convolutional Neural Network
Gradient-based Learning Applied to Document Recognition
Data Mining, Neural Network and Genetic Programming
Data Mining, Neural Network and Genetic Programming
A VERY Brief Introduction to Convolutional Neural Network using TensorFlow 李 弘
Lecture 5 Smaller Network: CNN
Non-linear classifiers Neural networks
Introduction to Neural Networks
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
Deep learning Introduction Classes of Deep Learning Networks
network of simple neuron-like computing elements
Convolutional neural networks Abin - Roozgard.
Neural Network - 2 Mayank Vatsa
CSC 578 Neural Networks and Deep Learning
On Convolutional Neural Network
Analysis of Trained CNN (Receptive Field & Weights of Network)
Convolutional Neural Networks
Face Recognition: A Convolutional Neural Network Approach
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
CSC 578 Neural Networks and Deep Learning
Outline Announcement Neural networks Perceptrons - continued
Presentation transcript:

1 Convolutional neural networks Abin - Roozgard

2  Introduction  Drawbacks of previous neural networks  Convolutional neural networks  LeNet 5  Comparison  Disadvantage  Application Presentation layout

3 introduction

4 Introduction  Convolutional neural networks  Signal processing, Image processing  improvement over the multilayer perceptron  performance, accuracy and some degree of invariance to distortions in the input images

5 Drawbacks

6 Structure Types of Decision Regions Exclusive-OR Problem Classes with Meshed regions Most General Region Shapes Single-Layer Two-Layer Three-Layer Half Plane Bounded By Hyper plane Convex Open Or Closed Regions Arbitrary (Complexity Limited by No. of Nodes) A AB B A AB B A AB B B A B A B A Behavior of multilayer neural networks

7 Multi-layer perceptron and image processing  One or more hidden layers  Sigmoid activations functions

8  the number of trainable parameters becomes extremely large Drawbacks of previous neural networks

9  Little or no invariance to shifting, scaling, and other forms of distortion Drawbacks of previous neural networks

10  Little or no invariance to shifting, scaling, and other forms of distortion Drawbacks of previous neural networks Shift left

11 Drawbacks of previous neural networks 154 input change from 2 shift left 77 : black to white 77 : white to black

12  scaling, and other forms of distortion Drawbacks of previous neural networks

13  the topology of the input data is completely ignored  work with raw data. Drawbacks of previous neural networks Feature 1 Feature 2

14 Drawbacks of previous neural networks Black and white patterns: Gray scale patterns : 32 * 32 input image

15 A A Drawbacks of previous neural networks

16 Improvement  Fully connected network of sufficient size can produce outputs that are invariant with respect to such variations.  Training time  Network size  Free parameters

17 Convolutional neural network (CNN)

18  In 1995, Yann LeCun and Yoshua Bengio introduced the concept of convolutional neural networks. History Yann LeCun, Professor of Computer Science The Courant Institute of Mathematical Sciences New York University Room 1220, 715 Broadway, New York, NY 10003, USA. (212)

19  CNN’s Were neurobiologically motivated by the findings of locally sensitive and orientation-selective nerve cells in the visual cortex.  They designed a network structure that implicitly extracts relevant features.  Convolutional Neural Networks are a special kind of multi-layer neural networks. About CNN’s

20  CNN is a feed-forward network that can extract topological properties from an image.  Like almost every other neural networks they are trained with a version of the back-propagation algorithm.  Convolutional Neural Networks are designed to recognize visual patterns directly from pixel images with minimal preprocessing.  They can recognize patterns with extreme variability (such as handwritten characters). About CNN’s

21  Classification  Convolutional neural network Classification Pre-processing for feature extraction Input f 1 … f n output f1f1 f 2 output Feature extraction Input Shift and distortion invariance classification

22 CNN’s Topology Feature extraction layer Convolution layer Shift and distortion invariance or Subsampling layer C S Feature maps

23 Feature extraction layer or Convolution layer features  detect the same feature at different positions in the input image.

24 Feature extraction w13 w12w11 w23w22w21 w33w32w Convolve withThreshold w13 w12w11 w23w22w21 w33w32w31

25 Feature extraction InputsCS  Shared weights: all neurons in a feature share the same weights (but not the biases).  In this way all neurons detect the same feature at different positions in the input image.  Reduce the number of free parameters.

26 Feature extraction  If a neuron in the feature map fires, this corresponds to a match with the template.

27 Subsampling layer  the subsampling layers reduce the spatial resolution of each feature map  By reducing the spatial resolution of the feature map, a certain degree of shift and distortion invariance is achieved.

28  the subsampling layers reduce the spatial resolution of each feature map Subsampling layer

29 Subsampling layer

30 Subsampling layer  The weight sharing is also applied in subsampling layers.

31  the weight sharing is also applied in subsampling layers  reduce the effect of noises and shift or distortion Feature map Subsampling layer

32 Up to now …

33 LeNet 5

34 LeNet5  Introduced by LeCun.  raw image of 32 × 32 pixels as input.

35 LeNet5  C1,C3,C5 : Convolutional layer.  5 × 5 Convolution matrix.  S2, S4 : Subsampling layer.  Subsampling by factor 2.  F6 : Fully connected layer.

36 LeNet5  All the units of the layers up to F6 have a sigmoidal activation function of the type:

37 0 F1 F2 F84 W1 W2 W84 Y0Y0 LeNet5 +1

38 LeNet5  About 187,000 connection.  About 14,000 trainable weight.

39 LeNet5

40 LeNet5

41 Comparison

42  Database : MNIST (60,000 handwritten digits )  Affine distortion : translation, rotation.  elastic deformations : corresponding to uncontrolled oscillations of the hand muscles.  MLP (this paper) : has 800 hidden unit. Comparison

43  “This paper” refer to reference[ 3] on references slide. Comparison

44 Disadvantages

45 Disadvantages  From a memory and capacity standpoint the CNN is not much bigger than a regular two layer network.  At runtime the convolution operations are computationally expensive and take up about 67% of the time.  CNN’s are about 3X slower than their fully connected equivalents (size-wise).

46 Disadvantages  Convolution operation  4 nested loops ( 2 loops on input image & 2 loops on kernel)  Small kernel size  make the inner loops very inefficient as they frequently JMP.  Cash unfriendly memory access  Back-propagation require both row-wise and column-wise access to the input and kernel image.  2-D Images represented in a row-wise-serialized order.  Column-wise access to data can result in a high rate of cash misses in memory subsystem.

47 Application

48 Application  A Convolutional neural network achieves 99.26% accuracy on a modified NIST database of hand-written digits.  MNIST database : Consist of 60,000 hand written digits uniformly distributed over 0-9. Mike O'Neill

49 Application

50 Application

51 Application

52 Application

53 References [1].Y. LeCun and Y. Bengio.“Convolutional networks for images, speech, and time-series.” In M. A. Arbib, editor, The Handbook of Brain Theory and Neural Networks. MIT Press, [2].Fabien Lauer, ChingY. Suen, Gérard Bloch,”A trainable feature extractor for handwritten digit recognition“, Elsevier, october [3].Patrice Y. Simard, Dave Steinkraus, John Platt, "Best Practices for Convolutional Neural Networks Applied to Visual Document Analysis," International Conference on Document Analysis and Recognition (ICDAR), IEEE Computer Society, Los Alamitos, pp , 2003.

54 Questions 8 4 6