 Karthik Gurumoorthy  Ajit Rajwade  Arunava Banerjee  Anand Rangarajan Department of CISE University of Florida 1.

Slides:



Advertisements
Similar presentations
Chapter 28 – Part II Matrix Operations. Gaussian elimination Gaussian elimination LU factorization LU factorization Gaussian elimination with partial.
Advertisements

Learning an Attribute Dictionary for Human Action Classification
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Structured Sparse Principal Component Analysis Reading Group Presenter: Peng Zhang Cognitive Radio Institute Friday, October 01, 2010 Authors: Rodolphe.
3D Reconstruction – Factorization Method Seong-Wook Joo KG-VISA 3/10/2004.
Learning sparse representations to restore, classify, and sense images and videos Guillermo Sapiro University of Minnesota Supported by NSF, NGA, NIH,
* * Joint work with Michal Aharon Freddy Bruckstein Michael Elad
A novel supervised feature extraction and classification framework for land cover recognition of the off-land scenario Yan Cui
1 Micha Feigin, Danny Feldman, Nir Sochen
Ilias Theodorakopoulos PhD Candidate
An Introduction to Sparse Coding, Sparse Sensing, and Optimization Speaker: Wei-Lun Chao Date: Nov. 23, 2011 DISP Lab, Graduate Institute of Communication.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Principal Component Analysis
Image Super-Resolution Using Sparse Representation By: Michael Elad Single Image Super-Resolution Using Sparse Representation Michael Elad The Computer.
Dimensionality Reduction Chapter 3 (Duda et al.) – Section 3.8
Principal Component Analysis
Sparse and Overcomplete Data Representation
Efficient Sparse Coding Algorithms
Image Denoising via Learned Dictionaries and Sparse Representations
CS 790Q Biometrics Face Recognition Using Dimensionality Reduction PCA and LDA M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive.
Approximate Nearest Subspace Search with Applications to Pattern Recognition Ronen Basri, Tal Hassner, Lihi Zelnik-Manor presented by Andrew Guillory and.
Real-time Computer Vision with Scanning N-Tuple Grids Simon Lucas Computer Science Dept.
New Results in Image Processing based on Sparse and Redundant Representations Michael Elad The Computer Science Department The Technion – Israel Institute.
Singular Value Decomposition COS 323. Underconstrained Least Squares What if you have fewer data points than parameters in your function?What if you have.
Face Recognition using PCA (Eigenfaces) and LDA (Fisherfaces)
Image Denoising with K-SVD Priyam Chatterjee EE 264 – Image Processing & Reconstruction Instructor : Prof. Peyman Milanfar Spring 2007.
Face Recognition Jeremy Wyatt.
Computing Sketches of Matrices Efficiently & (Privacy Preserving) Data Mining Petros Drineas Rensselaer Polytechnic Institute (joint.
Previously Two view geometry: epipolar geometry Stereo vision: 3D reconstruction epipolar lines Baseline O O’ epipolar plane.
Recent Trends in Signal Representations and Their Role in Image Processing Michael Elad The CS Department The Technion – Israel Institute of technology.
3D Geometry for Computer Graphics
Biomedical Image Analysis and Machine Learning BMI 731 Winter 2005 Kun Huang Department of Biomedical Informatics Ohio State University.
SVD(Singular Value Decomposition) and Its Applications
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
Under the guidance of Dr. K R. Rao Ramsanjeev Thota( )
Face Recognition and Feature Subspaces
Feature extraction 1.Introduction 2.T-test 3.Signal Noise Ratio (SNR) 4.Linear Correlation Coefficient (LCC) 5.Principle component analysis (PCA) 6.Linear.
Cs: compressed sensing
Fast and incoherent dictionary learning algorithms with application to fMRI Authors: Vahid Abolghasemi Saideh Ferdowsi Saeid Sanei. Journal of Signal Processing.
SVD: Singular Value Decomposition
Yang, Luyu.  Postal service for sorting mails by the postal code written on the envelop  Bank system for processing checks by reading the amount of.
Learning to Sense Sparse Signals: Simultaneous Sensing Matrix and Sparsifying Dictionary Optimization Julio Martin Duarte-Carvajalino, and Guillermo Sapiro.
CSE 185 Introduction to Computer Vision Face Recognition.
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
CSE 446 Dimensionality Reduction and PCA Winter 2012 Slides adapted from Carlos Guestrin & Luke Zettlemoyer.
PCA vs ICA vs LDA. How to represent images? Why representation methods are needed?? –Curse of dimensionality – width x height x channels –Noise reduction.
Quiz Week 8 Topical. Topical Quiz (Section 2) What is the difference between Computer Vision and Computer Graphics What is the difference between Computer.
2D-LDA: A statistical linear discriminant analysis for image matrix
Matrix Factorization & Singular Value Decomposition Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Martina Uray Heinz Mayer Joanneum Research Graz Institute of Digital Image Processing Horst Bischof Graz University of Technology Institute for Computer.
From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images Alfred M. Bruckstein (Technion), David L. Donoho (Stanford), Michael.
Face Recognition based on 2D-PCA and CNN
Jeremy Watt and Aggelos Katsaggelos Northwestern University
Lecture 8:Eigenfaces and Shared Features
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
PCA vs ICA vs LDA.
Singular Value Decomposition
Mean transform , a tutorial
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
Lecture 21 SVD and Latent Semantic Indexing and Dimensional Reduction
Rob Fergus Computer Vision
Domingo Mery Department of Computer Science
Outline H. Murase, and S. K. Nayar, “Visual learning and recognition of 3-D objects from appearance,” International Journal of Computer Vision, vol. 14,
A Digital Watermarking Scheme Based on Singular Value Decomposition
Image Coding and Compression
Image Compression via SVD
* * Joint work with Michal Aharon Freddy Bruckstein Michael Elad
Learning-Based Low-Rank Approximations
Presentation transcript:

 Karthik Gurumoorthy  Ajit Rajwade  Arunava Banerjee  Anand Rangarajan Department of CISE University of Florida 1

 A new approach to lossy image compression based on machine learning.  Key idea: Learning of Matrix Ortho-normal Bases from training data to efficiently code images.  Applied to compression of well-known face databases like ORL, Yale.  Competitive with JPEG. 2

Vector Conventional learning methods in vision like PCA, ICA, etc. Image 3

Our approach following Rangarajan [EMMCVPR-2001] & Ye [JMLR-2004] Treated as a Image Matrix 4

Image of size divided into N patches of size each treated as a Matrix. Image 5

6 = PUSV U and V: Ortho-normal matrices S: Diagonal Matrix of singular values

7 useful for compression (e.g.: SSVD [Ranade et al- IVC 2007]).

8  Consider a set of N image patches:  SVD of each patch gives:  Costly in terms of storage as we need to store N ortho-normal basis pairs.

 Produce ortho-normal basis-pairs, common for all N patches.  Since storing the basis pairs is not expensive. 9

10 Non-diagonal Non-sparse

 What sparse matrix will optimally reconstruct from ?  Optimally = least error:  Sparse = matrix has at most some non-zero elements. 11

 We have a simple, provably optimal greedy method to compute such a 1. Compute the matrix. 2. In matrix, nullify all except the largest elements to produce. 12

 A set of N image patches.  Learning K << N ortho-normal basis pairs 13 Memberships Projection Matrices

 Input: N image patches of size.  Output: K pairs of ortho-normal bases called as dictionary. 14

 Divide each test image into patches of size  Fix per-pixel average error (say e), similar to the “quality” user-parameter in JPEG. 15

RPP = number of bits per pixel 17

bits 0.92 bits1.36 bits 1.78 bits3.023 bits

 Size of original database is 3.46 MB.  Size of dictionary of 50 ortho-normal basis pairs is 56 KB=0.05MB.  Size of database after compression and coding with our method with e = is 1.3 MB.  Total compression rate achieved is 61%. 19

RPP = number of bits per pixel 20

 New lossy image compression method using machine learning.  Key idea 1: matrix based image representation.  Key idea2: Learning small set of matrix ortho- normal basis pairs tuned to a database.  Results competitive with JPEG standard.  Future extensions: video compression. 21

 A. Rangarajan, Learning matrix space image representations, Energy Minimizing Methods in Computer Vision and Pattern Recognition,  J. Ye, Generalized low rank approximation of matrices, Journal of Machine Learning Research,2004.  M. Aharon, M. Elad and A. Bruckstein, The K-SVD: An algorithm for designing of overcomplete dictionaries for sparse representation. IEEE Transactions on Signal Processing,  A. Ranade, S. Mahabalarao and S. Kale. A variation on SVD based image compression. Image and Vision Computing,

23