AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ.

Slides:



Advertisements
Similar presentations
EigenFaces and EigenPatches Useful model of variation in a region –Region must be fixed shape (eg rectangle) Developed for face recognition Generalised.
Advertisements

3D Geometry for Computer Graphics
Zhimin CaoThe Chinese University of Hong Kong Qi YinITCS, Tsinghua University Xiaoou TangShenzhen Institutes of Advanced Technology Chinese Academy of.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Graph Embedding and Extensions: A General Framework for Dimensionality Reduction Keywords: Dimensionality reduction, manifold learning, subspace learning,
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
São Paulo Advanced School of Computing (SP-ASC’10). São Paulo, Brazil, July 12-17, 2010 Looking at People Using Partial Least Squares William Robson Schwartz.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
RETRIEVAL OF MULTIMEDIA OBJECTS USING COLOR SEGMENTATION AND DIMENSION REDUCTION OF FEATURES Mingming Lu, Qiyu Zhang, Wei-Hung Cheng, Cheng-Chang Lu Department.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Smart Traveller with Visual Translator for OCR and Face Recognition LYU0203 FYP.
Atul Singh Junior Undergraduate CSE, IIT Kanpur.  Dimension reduction is a technique which is used to represent a high dimensional data in a more compact.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Lightseminar: Learned Representation in AI An Introduction to Locally Linear Embedding Lawrence K. Saul Sam T. Roweis presented by Chan-Su Lee.
Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul Reference: "Nonlinear dimensionality reduction by locally.
CS 485/685 Computer Vision Face Recognition Using Principal Components Analysis (PCA) M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
Linear Algebra and Image Processing
SVD(Singular Value Decomposition) and Its Applications
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Summarized by Soo-Jin Kim
Dimensionality Reduction: Principal Components Analysis Optional Reading: Smith, A Tutorial on Principal Components Analysis (linked to class webpage)
1 Graph Embedding (GE) & Marginal Fisher Analysis (MFA) 吳沛勳 劉冠成 韓仁智
General Tensor Discriminant Analysis and Gabor Features for Gait Recognition by D. Tao, X. Li, and J. Maybank, TPAMI 2007 Presented by Iulian Pruteanu.
1 Recognition by Appearance Appearance-based recognition is a competing paradigm to features and alignment. No features are extracted! Images are represented.
IEEE TRANSSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
Data Reduction. 1.Overview 2.The Curse of Dimensionality 3.Data Sampling 4.Binning and Reduction of Cardinality.
Classification Course web page: vision.cis.udel.edu/~cv May 12, 2003  Lecture 33.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
A Two-level Pose Estimation Framework Using Majority Voting of Gabor Wavelets and Bunch Graph Analysis J. Wu, J. M. Pedersen, D. Putthividhya, D. Norgaard,
Face Recognition: An Introduction
ISOMAP TRACKING WITH PARTICLE FILTER Presented by Nikhil Rane.
Levels of Image Data Representation 4.2. Traditional Image Data Structures 4.3. Hierarchical Data Structures Chapter 4 – Data structures for.
Jan Kamenický.  Many features ⇒ many dimensions  Dimensionality reduction ◦ Feature extraction (useful representation) ◦ Classification ◦ Visualization.
Project by: Cirill Aizenberg, Dima Altshuler Supervisor: Erez Berkovich.
EE4-62 MLCV Lecture Face Recognition – Subspace/Manifold Learning Tae-Kyun Kim 1 EE4-62 MLCV.
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
Timo Ahonen, Abdenour Hadid, and Matti Pietikainen
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
MACHINE LEARNING 7. Dimensionality Reduction. Dimensionality of input Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Math 285 Project Diffusion Maps Xiaoyan Chong Department of Mathematics and Statistics San Jose State University.
Feature Extraction 主講人:虞台文. Content Principal Component Analysis (PCA) PCA Calculation — for Fewer-Sample Case Factor Analysis Fisher’s Linear Discriminant.
2D-LDA: A statistical linear discriminant analysis for image matrix
Intro. ANN & Fuzzy Systems Lecture 16. Classification (II): Practical Considerations.
Feature Extraction 主講人:虞台文.
Unsupervised Learning II Feature Extraction
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Spectral Methods for Dimensionality
Principal Component Analysis (PCA)
LECTURE 10: DISCRIMINANT ANALYSIS
Unsupervised Riemannian Clustering of Probability Density Functions
PRESENTED BY Yang Jiao Timo Ahonen, Matti Pietikainen
Recognition using Nearest Neighbor (or kNN)
Machine Learning Dimensionality Reduction
PCA vs ICA vs LDA.
Outline Nonlinear Dimension Reduction Brief introduction Isomap LLE
Object Modeling with Layers
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
Computer Vision Lecture 16: Texture II
Cheng-Yi, Chuang (莊成毅), b99
Feature space tansformation methods
LECTURE 09: DISCRIMINANT ANALYSIS
NonLinear Dimensionality Reduction or Unfolding Manifolds
Lecture 16. Classification (II): Practical Considerations
Presentation transcript:

AGE ESTIMATION: A CLASSIFICATION PROBLEM HANDE ALEMDAR, BERNA ALTINEL, NEŞE ALYÜZ, SERHAN DANİŞ

Project Overview

Subset Overview Aging Subset of Bosphorus Database: – 1-4 neutral and frontal 2D images of subjects – 105 subjects – Total of 298 scans – Age range: [18-60] – Age distribution non uniform: average = 29.9

Project Overview Aging images of individuals is not present Aim: Age Estimation based on Age Classes 3 Classes: Age 96 samples samples Age>36 -> 41 samples

Preprocessing Registration Cropping Histogram Equalization Resizing

SUBSPACE ANALYSIS FOR AGE ESTIMATION Neşe Alyüz

Age Manifold Instead of learning a subject-specific aging pattern, a common aging trend can be learned Manifold embedding technique to learn the low-dimensional aging trend. Image space: Labels: Low-dim. representation: d<<D Mapping:

Orthogonal Locality Preserving Projections - OLPP Subspace learning technique Produces orthogonal basis functions on LPP LPP: The essential manifold structure preserved by measuring local neighborhood distances OLPP vs. PCA for age manifold: OLPP is supervised, PCA is unsupervised OLPP better, since age labeling is used for learning X Size of training data for OLPP should be LARGE enough

Locality Preserving Projection - LPP aka: Laplacianface Approach Linear dimensionality reduction algorithm Builds a graph: based on neighborhood information Obtains a linear transformation: Neighborhood information is preserved

LPP S: similarity matrix defined on data points (weights) L = D – S : graph Laplacian D: diagonal sum matrix of S measures local density around a sample point Minimization problem: with the constraint : => Minimizing this function: ensure that if x i and x j are close then their projections y i and y j are also close

LPP Generalized eigenvalue problem: Basis functions are the eigenvectors of: Not symmetric, therefore the basis functions are not orthogonal

OLPP In LPP, basis functions are nonorthogonal – > reconstruction is difficult OLPP produces orthogonal basis functions – > has more locality preserving power

OLPP – Algorithmic Outline (1) Preprocessing: PCA projection (2) Constructing the Adjacency Graph (3) Choosing the Locality Weights (4) Computing the Orthogonal Basis Functions (5) OLPP Embedding

(1) Preprocessing: PCA Prjection XDX T can be singular To overcome the singularity problem -> PCA Throwing away components, whose corresponding eigenvalues are zero. Transformation matrix: W PCA Extracted features become statistically uncorrelated

(2) Constructing The Adjacency Graph G: a graph with n nodes If face images x i and x j are connected (has the same label) then an edge exists in-between.

(3) Choosing the Locality Weights S: weight matrix If node i and j are connected: Weights: heat kernel function Models the local structure of the manifold

(4) Computing the Orthogonal Basis Functions D: diagonal matrix, column sum of S L : laplacian matrix, L = D – S Orthogonal basis vectors: Two extra matrices defined: Computing the basis vectors: – Compute a 1 : eigenvector of with the greatest eigenvalue – Compute a k : eigenvector of with the greatest eigenvalue

(5) OLPP Embedding Let: Overall embedding:

Subspace Methods: PCA vs. OLPP Face Recognition Results on ORL

Subspace Methods: PCA vs. OLPP Face Recognition Results on Aging Subset of the Bosphorus Database Age Estimation (Classification) Results on Aging Subset of the Bosphorus Database

FEATURE EXTRACTION: LOCAL BINARY PATTERNS Hande Alemdar

Feature Extraction LBP - Local Binary Patterns

Local Binary Patterns More formally For 3x3 neighborhood we have 256 patterns Feature vector size = 256 where

Uniform LBP Uniform patterns can be used to reduce the length of the feature vector and implement a simple rotation-invariant descriptor If the binary pattern contains at most two bitwise transitions from 0 to 1 or vice versa when the bit pattern is traversed circularly  Uniform – is uniform – (2 transitions) – (2 transitions) For 3x3 neighborhood we have 58 uniform patterns Feature vector size = 59

FEATURE EXTRACTION: GABOR FILTERING Serhan Daniş

Gabor Filter Band-pass filters used for feature extraction, texture analysis and stereo disparity estimation. Can be designed for a number of dilations and rotations. The filters with various dilations and rotations are convolved with the signal, resulting in a so-called Gabor space. This process is closely related to processes in the primary visual cortex.

Gabor Filter  A set of Gabor filters with different frequencies and orientations may be helpful for extracting useful features from an image.  We used 6 different rotations and 4 different scales on 16 overlapping patches of the images.  We generate 768 features for each image.

CLASSIFICATION Berna Altınel

EXPERIMENTAL DATASETS 1. FEATURES_50_45(LBP) 2. FEATURES_100_90(LBP) 3. FEATURES_ORIG(LBP) 4. FEATURES_50_45(GABOR) 5. FEATURES_100_90 (GABOR)

Estimate age, just based on the average value of the training set Experiment #1

EXPERIMENTAL RESULTS: INPUTMETHODMean(MRE) %Number of Correct classifications Number of miss classifications Features_ 50_45 -LBP Estimating the Average Age / / 298 Features_ 100_90 -LBP Features_orig -LBP Features_ 50_45 (GABOR) Features_ 100_90 (GABOR)

K-NEAREST-NEIGHBOR ALGORITHM Experiments #2 The K-nearest-neighbor (KNN) algorithm measures the distance between a query scenario and a set of scenarios in the data set.

EXPERIMENTAL RESULTS: INPUTFeatures_ 50_45 (LBP) Features_ 100_90 (LBP) Features_ orig (LBP) Features_ 50_45 (GABOR) Features_ 100_90 (GABOR) METHOD kNN-1(Euc Dist) MRE(%):5.05MRE(%):4.14MRE(%): MRE(%):3.88MRE(%):3.75 kNN-2(Euc Dist) MRE(%):6.77MRE(%):5.17MRE(%): MRE(%):4.92MRE(%):5.08 kNN-3(Euc Dist) MRE(%):7.50MRE(%):6.06MRE(%): MRE(%):5.79MRE(%):5.96 kNN-5(Euc Dist) MRE(%): MRE(%):9.36MRE(%): MRE(%): MRE(%): kNN-10(Euc Dist) MRE(%): MRE(%): MRE(%): MRE(%): MRE(%): kNN-15(Euc Dist) MRE(%): MRE(%): MRE(%): MRE(%): MRE(%):14.9 8

Features_ 50_45 (LBP) Features_ 100_90 (LBP) Features_ orig (LBP) Features_ 50_45 (GABOR) Features_ 100_90 (GABOR) METHOD AverageMRE(%): MRE(%): MRE(%): MRE(%): MRE(%): MissClass: 35 / 298 MissClass: 32 / 298 MissClass: 31 / 298 MissClass: 32 / 298 MissClass: 27 / 298 CorrectClas s:262 / 298 CorrectClas s: 265 / 298 CorrectClas s: 266 / 298 CorrectClas s:265 / 298 CorrectClas s:270 / 298

IN PROGRESS: 1. Parametric Classification 2. Mahalanobis distance can be used as the distance measure in kNN. [ 2 [ 2 [ 2 [

1. Other distance functions can be analyzed for kNN: 2. Normalization can be applied: POSSIBLE FUTURE WORK ITEMS: