A Comparative Study of Convolutional Neural Network Models with Rosenblatt’s Brain Model Abu Kamruzzaman, Atik Khatri , Milind Ikke, Damiano Mastrandrea,

Slides:



Advertisements
Similar presentations
Face Recognition: A Convolutional Neural Network Approach
Advertisements

Lecture 14 – Neural Networks
Neural Networks Lab 5. What Is Neural Networks? Neural networks are composed of simple elements( Neurons) operating in parallel. Neural networks are composed.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Skeleton Based Action Recognition with Convolutional Neural Network
Convolutional LSTM Networks for Subcellular Localization of Proteins
Philipp Gysel ECE Department University of California, Davis
1 Convolutional neural networks Abin - Roozgard. 2  Introduction  Drawbacks of previous neural networks  Convolutional neural networks  LeNet 5 
Recent developments in object detection
Faster R-CNN – Concepts
The Relationship between Deep Learning and Brain Function
Object Detection based on Segment Masks
Data Mining, Neural Network and Genetic Programming
Convolutional Neural Fabrics by Shreyas Saxena, Jakob Verbeek
Recurrent Neural Networks for Natural Language Processing
A Hierarchical Model of Reviews for Aspect-based Sentiment Analysis
Combining CNN with RNN for scene labeling (segmentation)
Ajita Rattani and Reza Derakhshani,
Inception and Residual Architecture in Deep Convolutional Networks
Intelligent Information System Lab
Lecture 5 Smaller Network: CNN
Efficient Deep Model for Monocular Road Segmentation
Deep Residual Learning for Image Recognition
Dynamic Routing Using Inter Capsule Routing Protocol Between Capsules
Deep Learning Convoluted Neural Networks Part 2 11/13/
Layer-wise Performance Bottleneck Analysis of Deep Neural Networks
Bird-species Recognition Using Convolutional Neural Network
A critical review of RNN for sequence learning Zachary C
Convolutional Neural Networks
Introduction to Neural Networks
Grid Long Short-Term Memory
Counting in Dense Crowds using Deep Learning
Deep Learning Hierarchical Representations for Image Steganalysis
Introduction to Deep Learning with Keras
network of simple neuron-like computing elements
Convolutional neural networks Abin - Roozgard.
CSC 578 Neural Networks and Deep Learning
Creating Data Representations
Optimization for Fully Connected Neural Network for FPGA application
A Proposal Defense On Deep Residual Network For Face Recognition Presented By SAGAR MISHRA MECE
Final Project presentation
On Convolutional Neural Network
Analysis of Trained CNN (Receptive Field & Weights of Network)
Designing Neural Network Architectures Using Reinforcement Learning
Logistic Regression & Transfer Learning
Deep Learning Some slides are from Prof. Andrew Ng of Stanford.
Word2Vec.
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Heterogeneous convolutional neural networks for visual recognition
Face Recognition: A Convolutional Neural Network Approach
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Model Compression Joseph E. Gonzalez
Department of Computer Science Ben-Gurion University of the Negev
Automatic Handwriting Generation
Car Damage Classification
Human-object interaction
Deep Object Co-Segmentation
Natalie Lang Tomer Malach
Image Processing and Multi-domain Translation
Neural Machine Translation using CNN
The experiments based on Recurrent Neural Networks
Learning and Memorization
End-to-End Facial Alignment and Recognition
CRCV REU 2019 Kara Schatz.
Real-time Object Recognition using deep learning-Raspberry Pi
Adrian E. Gonzalez , David Parra Department of Computer Science
Presented By: Firas Gerges (fg92)
Convolutional neural network based Alzheimer’s disease classification from magnetic resonance brain images RachnaJain NikitaJainaAkshayAggarwal D.
Presentation transcript:

A Comparative Study of Convolutional Neural Network Models with Rosenblatt’s Brain Model Abu Kamruzzaman, Atik Khatri , Milind Ikke, Damiano Mastrandrea, Poorva Shelke and Charles C. Tappert

Outline Introduction Project Outline Methodology Experiment Results Conclusion

What is Neural Networks A neural network is made up of neurons, which are a set of connected processors. Large number of processors are arranged in parallel and they work together Each tier receives output from the previous tier The last layer of the Neural Network produces the output.

Convolutional Neural network Works good with Image dataset. They learn more easily and they learn the things that we want They try to recognize the components of the image and thus predict the larger structure.

Pre-Trained Models

Pre-Trained Model Usage: Saves time A pre-trained model has been previously trained on a dataset and contains the weights and biases that represent the features of the dataset it was trained. Learned features are often transferable to different data. Usage: Saves time It is a model developed by another entity that is reused for the same purpose. Developers can use existing model as a starting point. The transferability between projects is more efficient.

VGG16 A study of convolutional neural networks It achieves 92.7% accuracy over the the top-5 test accuracy in ImageNet Dataset.

ResNet50 ResNet is short name for Residual Networks. It introduces Residual Learning Nothing but subtraction of feature learned from input of the layer

MobileNets MobileNets are small, low-latency, low-powered models They can be build upon classification, detections, embedding and segmentation MobileNets split the convolution into 3x3 depth wise conv and 1x1 pointwise conv.

InceptionV3 InceptionV3 is a variant of InceptionV2 with an addition of BN-Auxiliary. BN-auxiliary refers to the version in which a fully connected layer of auxiliary classified is also-normalized, not just convolutions.

Purpose and Methodology ! Compare, discuss, and experiment the CNN models with the proposed Memory model and discuss the limitations of the CNN A sequential memory model, which stores the results of the hidden layer This storage (like a memory) can be used in the future predictions Our experiments and results focuses on: How well CNN classify the spatial data ? How CNN learned hidden representation, does it store the hidden layers for future predictions

Experiments - CNN models

Pre-trained Input Images: Experiment Setup Pre-trained Input Images: Python version 3.0 Knowledge of Pre-trained Models Weights has been downloaded in Keras and loaded for the experiment on the images

Experiment 1 – ResNet50 model Experiment 2 – VGG16 model

Experiment 3 – InceptionV3 model Experiment 4 – MobileNet model Experiment 3 – InceptionV3 model

Results After conducting the experiment (from 1-4), we found the following results Those CNN models were able to classify the images with a very good accuracy

CNN Pre-trained Model Accuracy Comparing the pre-trained model accuracy Pre-trained models Image 1: Banana Image 2: Orange Image 3: Lion Image 4: Seashore ResNet50 0.989 0.998 0.999 0.375 VGG16 0.986 0.982 0.594 InceptionV3 1.0 0.992 0.864 MobileNet 0.988 0.426

Experiment-5 MNIST digits dataset

Results After conducting an experiment on MNIST data, it can be found that, CNN first extract all the features from the images Then, considering those features it transforms the features with series of operations (mainly Convolution operation and MaxPooling operation) Undertaking all the layers, the final layer is just a high level layer representation which has learned discrete patterns of features without considering or storing any past sequences of patterns.

Conclusion CNN is mostly used for classification of Spatial data CNN does not store or remember past sequences of patterns that can be used for future predictions The Proposed architecture (like Rosenblatt Memory model) require more research and exploration to solve problems related to speech recognition, language modelling, language translation, and audio/video processing, where the past context and sequences are important for the future predictions

Questions ? Thank You