A brief review of non-neural-network approaches to deep learning

Slides:



Advertisements
Similar presentations
Greedy Layer-Wise Training of Deep Networks
Advertisements

Neural networks Introduction Fitting neural networks
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
ImageNet Classification with Deep Convolutional Neural Networks
Lecture 14 – Neural Networks
Supervised and Unsupervised learning and application to Neuroscience Cours CA6b-4.
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
Machine learning Image source:
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
What is the Best Multi-Stage Architecture for Object Recognition Kevin Jarrett, Koray Kavukcuoglu, Marc’ Aurelio Ranzato and Yann LeCun Presented by Lingbo.
Machine learning Image source:
A Genetic Algorithms Approach to Feature Subset Selection Problem by Hasan Doğu TAŞKIRAN CS 550 – Machine Learning Workshop Department of Computer Engineering.
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
A shallow introduction to Deep Learning
M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
1/18 New Feature Presentation of Transition Probability Matrix for Image Tampering Detection Luyi Chen 1 Shilin Wang 2 Shenghong Li 1 Jianhua Li 1 1 Department.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Introduction to Deep Learning
Convolutional Neural Network
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Deep Learning and Deep Reinforcement Learning. Topics 1.Deep learning with convolutional neural networks 2.Learning to play Atari video games with Deep.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Neural networks and support vector machines
Big data classification using neural network
Convolutional Sequence to Sequence Learning
Automatic Lung Cancer Diagnosis from CT Scans (Week 1)
Faster R-CNN – Concepts
Deep Feedforward Networks
Summary of “Efficient Deep Learning for Stereo Matching”
Compact Bilinear Pooling
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
ECE 5424: Introduction to Machine Learning
Computer Science and Engineering, Seoul National University
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Matt Gormley Lecture 16 October 24, 2016
Presenter: Chu-Song Chen
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
Intelligent Information System Lab
Training Techniques for Deep Neural Networks
NEURAL NETWORK APPROACHES FOR AUTOMOBILE MPG PREDICTION
Understanding the Difficulty of Training Deep Feedforward Neural Networks Qiyue Wang Oct 27, 2017.
CS6890 Deep Learning Weizhen Cai
Dynamic Routing Using Inter Capsule Routing Protocol Between Capsules
By: Kevin Yu Ph.D. in Computer Engineering
Introduction to Neural Networks
Grid Long Short-Term Memory
Logistic Regression & Parallel SGD
ECE 599/692 – Deep Learning Lecture 9 – Autoencoder (AE)
Object Classification through Deconvolutional Neural Networks
Neural Networks Geoff Hulten.
Lecture: Deep Convolutional Neural Networks
Graph Neural Networks Amog Kamsetty January 30, 2019.
Artificial Intelligence 10. Neural Networks
实习生汇报 ——北邮 张安迪.
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Reuben Feinman Research advised by Brenden Lake
Department of Computer Science Ben-Gurion University of the Negev
Automatic Handwriting Generation
Introduction to Neural Networks
CS295: Modern Systems: Application Case Study Neural Network Accelerator Sang-Woo Jun Spring 2019 Many slides adapted from Hyoukjun Kwon‘s Gatech “Designing.
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION
Learning and Memorization
Presentation transcript:

A brief review of non-neural-network approaches to deep learning Naiyan Wang

Outline Non-NN Approaches Discussion Deep Convex Net Extreme Learning Machine PCAnet Deep Fisher Net (Already presented before) Discussion

Deep convex net Each module is a two- layer convex network. After we get the prediction from each module, we concatenate it with the original input, and send it to a new module.

Deep Convex Net For each module We minimize U has a closed form solution: Learning of W relies on gradient descent: Note that no global fine tuning involved, so it can stack up to more than 10 layers. (Fast Training!)

Deep Convex Net A bit wired of why this works. The learned features in mid-layers are NOT representative for the input. Maybe learn the correlation between prediction and input could help? Discussion?

Deep Convex Net

Deep Convex Net

Extreme Learning Machine It is also a two layer networks: The first layer performs random projection of input data. The second layer performs OLS/Ridge regression to learn the weight. After that, we could take the transpose of the learned weight as the projection matrix, and stack several ELM into a deep one.

Extreme Learning Machine Extremely fast learning Note that even with simple random projection and linear transformation, the results still can be improved!

PCANet In the first two layers, use patch PCA to learn the filters. Then it binarizes the output in second layer, and calculate the histogram within a block.

PCANet To learn the filters, the authors also proposed to use Random initialization and LDA. The results are acceptable in a wide range of datasets.

Summary Most of the paper (except deep Fisher Net) report their results on relatively toy data. We cannot draw any conclusion about their performance. This could enlighten us some possible research directions.

Discussion Why deep architectures always help? (We don’t concern about overfitting now) The representation power increases exponentially as more layers add in. However the number of parameters increases linearly as more layers add in. Given a fixed budget, this is a better way to organize the model. Take PCA net as an example, if there are m, n neurons at first and second layer, then there exists an equivalent m*n single layer net.

Discussion Why CNN is so successful in image classification? Data abstraction Locality! (The image is a 2D structure with strong local correlation.) The convolution architecture could propagate local information to a broader region 1st: m * m, 2nd : n * n, then it corresponds to (m + n - 1) * (m + n - 1) in the original image. This advantage is further expanded by spatial pooling. Other ways to concern about these two issues simultaneously?

Discussion Convolution is a dense architecture. It induces a lot of unnecessary computation. Could we come up a greedy or more clever selection in each layer to just focus on those discriminative patches? Or possibly a “convolutional cascade”?

Discussion Random weights are adopted several times, and it yields acceptable results. Pros: Data independent Fast Cons: Data independent  So could we combine random weights and learned weights to combat against overfitting? Some work have been done on combining deterministic NN and stochastic NN.