4.2 Data Input-Output Representation

Slides:



Advertisements
Similar presentations
Aggregating local image descriptors into compact codes
Advertisements

Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
11/11/02 IDR Workshop Dealing With Location Uncertainty in Images Hasan F. Ates Princeton University 11/11/02.
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
嵌入式視覺 Feature Extraction
Designing Facial Animation For Speaking Persian Language Hadi Rahimzadeh June 2005.
電腦視覺 Computer and Robot Vision I Chapter2: Binary Machine Vision: Thresholding and Segmentation Instructor: Shih-Shinh Huang 1.
Quadtrees, Octrees and their Applications in Digital Image Processing
Unsupervised Learning With Neural Nets Deep Learning and Neural Nets Spring 2015.
Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy.
Pattern Recognition Topic 1: Principle Component Analysis Shapiro chap
Quadtrees, Octrees and their Applications in Digital Image Processing
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Oral Defense by Sunny Tang 15 Aug 2003
Dan Simon Cleveland State University
Radial-Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Presented by: Kamakhaya Argulewar Guided by: Prof. Shweta V. Jain
Presented by Tienwei Tsai July, 2005
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
S EGMENTATION FOR H ANDWRITTEN D OCUMENTS Omar Alaql Fab. 20, 2014.
Classification / Regression Neural Networks 2
1 Pattern Classification X. 2 Content General Method K Nearest Neighbors Decision Trees Nerual Networks.
Quadtrees, Octrees and their Applications in Digital Image Processing.
Handwritten Recognition with Neural Network Chatklaw Jareanpon, Olarik Surinta Mahasarakham University.
CS654: Digital Image Analysis
Neural Networks Lecture 4 out of 4. Practical Considerations Input Architecture Output.
Optical Character Recognition
1. 2 What is Digital Image Processing? The term image refers to a two-dimensional light intensity function f(x,y), where x and y denote spatial(plane)
Data statistics and transformation revision Michael J. Watts
Sparse Coding: A Deep Learning using Unlabeled Data for High - Level Representation Dr.G.M.Nasira R. Vidya R. P. Jaia Priyankka.
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Data Transformation: Normalization
Recognition of biological cells – development
INTRODUCTION TO GEOGRAPHICAL INFORMATION SYSTEM
LECTURE 11: Advanced Discriminant Analysis
Data Mining, Neural Network and Genetic Programming
Instance Based Learning
Artificial Intelligence (CS 370D)
Recognizing Deformable Shapes
Image Recognition. Contents: Motivation Objective Definition Introduction Preprocessing / Edge Detection Neural Networks in Image Recognition Practical.
Mean Shift Segmentation
Recognition using Nearest Neighbor (or kNN)
Final Year Project Presentation --- Magic Paint Face
Object Recognition in the Dynamic Link Architecture
Random walk initialization for training very deep feedforward networks
Classification / Regression Neural Networks 2
Computer Vision Lecture 4: Color
School of Computer Science & Engineering
Principal Component Analysis
PCA vs ICA vs LDA.
In summary C1={skin} C2={~skin} Given x=[R,G,B], is it skin or ~skin?
Image Segmentation Techniques
Computer Vision Lecture 16: Texture II
Zhengjun Pan and Hamid Bolouri Department of Computer Science
Digital Image Fundamentals
Creating Data Representations
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Grey Level Enhancement
Sparseness and Expansion in Sensory Representations
Neural networks (3) Regularization Autoencoder
CIS 4350 Image ENHANCEMENT SPATIAL DOMAIN
Fourier Transform of Boundaries
An introduction to: Deep Learning aka or related to Deep Neural Networks Deep Structural Learning Deep Belief Networks etc,
Anti-Faces for Detection
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Morphological Filters Applications and Extension Morphological Filters
Presentation transcript:

4.2 Data Input-Output Representation Background ( Ref: Handbook of Neural Computation ) Saund: Key theme in AI is to discover good rep. for the problem at hand. A good rep. makes explicit info. useful to the computation, it strips away obscuring clutter, it reduces info. to its essentials. Raw Data Transf. NN Final Output Input Output Data Rep. is more important/critical than network topology Goals of Data Rep. [Garbage In Garbage Out] Feature Enhancement / Data Reduction for Separability Similar (Diff) events  Similar (Diff) rep. for better interpolation More elements for important features Compactness without losing (or even enhancing) info. for fast learning Preserve feature info. (clustering / metric info) Ex. Binary coding could destroy metric 01111  10000 [Hamming metric]

Ex. Character Recognition Raw data = 5 64bit binary vectors. Representation ① raw data ② any code for five characters ③ shape features : horizontal and vertical spars, their ratio, relative positions of the spars Other extreme: Wasserman : raw data may be more useful in cases where essential features are unknown. NN can discover features in hidden neurons.

2. Data Preprocessing techniques Data sets are plagued by Noise, Bias, Large variations in dynamic range . . . (1) Normalize  Remove large dynamic variances over one or more dimensions in data Ex. ① Normalize gray scale image  invariant to light condition ② Normalize speech signal  invariant to absolute volume level ③ Normalize with respect to position and size - Character recognition

(3) Principal Component Analysis – Dim. Reduction (2) Nomalization Algorithms : ① One Way to embed magnitude ( l ) info. :  Normalize ② Row Norm. (2-D) : for each row, divide by the mean value ③ Column Norm. (2-D) (3) Principal Component Analysis – Dim. Reduction

3. Case Study : Face Recognition (profile) Data Reduction 416 x 320  16x2  23 (Discard High freq. components) (= 133,120 pixels) ① Efficient tech. to extract high interest features. ② Method for data reduction with minimal info. loss ③ DCT applied to reduced vector descriptions enhances info. content, provides invariance to small change and increased separability.

1 0 0 0 0 0 0 0 0 1 1 1 1 0 0 0 0 0 0 1 4. Actual Coding Schemes (1) Local R O Y G B I V (2) Distributed C2 C1 C0 1 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 1 1 1 1 In Local representation, each node may be [0 1] [-1 1] or even continuous when more than one node can be active to indicate presence of two or more features. (3) Coarse Distributed Wider overlapping receptive fields 36 nodes 27 nodes

Students’ Questions from 2005: While DCT in facial image processing helps data reduction, does it also help face recognition ? Since Human faces are all alike, its Fourier transform will also be similar. Spatial features will be more relevant to recognition. Normalization or Coding will reduce data or help classification. But, isn’t the process going to delay the overall learning time ? Coarse distributed coding will reduce the total number of nodes. However, when a single node is represented in overlapped fashion, isn’t the additional info. needed such as the overlap position, etc. ?

When an NN technique is used for character or speech recognition, how does its performance compare with non-NN approaches ? NN can be applied to many problems. Any application where NN is hard to apply to ? Is there any general measure to tell the importance of information in feature extraction ? If line search is used to find an optimal learning rate, the number of steps may decrease but I am afraid the overall processing time may increase. Can better separation for classification result via data representation ? Can the information content increase via a good data rep. ?

5. Discrete Coding (1) Simple Sum (fault tolerance, but requires many nodes for large numbers) 5 = 000011111 = 110000111 = (2) Value Unit Encoding 1-3 4-6 7-9 10-12 13-15 2 = 1-3 4-6 7-9 10-12 13-15 10 = (3) Discrete Thermometer x >0 x>3 x>6 x>9 x>12 2 = 10 = x >0 x>3 x>6 x>9 x>12

6. Continuous Coding (1) Simple Analog • For an activation ai range of [0,1] or [-1,1], value in range [u, v] = (v - u) ai + u • Logarithmic scale for data set with large dynamic range (2) Continuous Thermometer

Ref. “Neural Network Perception For Mobile Robot Guidance,” (3) Proportional Coarse Coding Pomerleau used a Gaussian Smearing function to represent steering directions in an Autonomous Land Vehicle In a Neural Network (ALVINN) Ref. “Neural Network Perception For Mobile Robot Guidance,” D. Pomerleau, Kluwer, 93. Slight Right Turn