Outlier Processing via L1-Principal Subspaces

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Yinyin Yuan and Chang-Tsun Li Computer Science Department
Aggregating local image descriptors into compact codes
Object Specific Compressed Sensing by minimizing a weighted L2-norm A. Mahalanobis.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Clustering: Introduction Adriano Joaquim de O Cruz ©2002 NCE/UFRJ
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Data preprocessing before classification In Kennedy et al.: “Solving data mining problems”
Robust Multi-Kernel Classification of Uncertain and Imbalanced Data
Principal Component Analysis
Reconstruction with Adaptive Feature-Specific Imaging Jun Ke 1 and Mark A. Neifeld 1,2 1 Department of Electrical and Computer Engineering, 2 College of.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
ON THE IMPROVEMENT OF IMAGE REGISTRATION FOR HIGH ACCURACY SUPER-RESOLUTION Michalis Vrigkas, Christophoros Nikou, Lisimachos P. Kondi University of Ioannina.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Unsupervised Learning
Visual Recognition Tutorial
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Facial Recognition CSE 391 Kris Lord.
Principal Component Analysis (PCA) for Clustering Gene Expression Data K. Y. Yeung and W. L. Ruzzo.
Graph-based consensus clustering for class discovery from gene expression data Zhiwen Yum, Hau-San Wong and Hongqiang Wang Bioinformatics, 2007.
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
HMM-BASED PSEUDO-CLEAN SPEECH SYNTHESIS FOR SPLICE ALGORITHM Jun Du, Yu Hu, Li-Rong Dai, Ren-Hua Wang Wen-Yi Chu Department of Computer Science & Information.
Summarized by Soo-Jin Kim
-1- ICA Based Blind Adaptive MAI Suppression in DS-CDMA Systems Malay Gupta and Balu Santhanam SPCOM Laboratory Department of E.C.E. The University of.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
EMIS 8381 – Spring Netflix and Your Next Movie Night Nonlinear Programming Ron Andrews EMIS 8381.
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
A Passive Approach to Sensor Network Localization Rahul Biswas and Sebastian Thrun International Conference on Intelligent Robots and Systems 2004 Presented.
CSE 185 Introduction to Computer Vision Face Recognition.
Sparse Signals Reconstruction Via Adaptive Iterative Greedy Algorithm Ahmed Aziz, Ahmed Salim, Walid Osamy Presenter : 張庭豪 International Journal of Computer.
Speech Lab, ECE, State University of New York at Binghamton  Classification accuracies of neural network (left) and MXL (right) classifiers with various.
Optimal Reverse Prediction: Linli Xu, Martha White and Dale Schuurmans ICML 2009, Best Overall Paper Honorable Mention A Unified Perspective on Supervised,
Irfan Ullah Department of Information and Communication Engineering Myongji university, Yongin, South Korea Copyright © solarlits.com.
Dimension reduction (1) Overview PCA Factor Analysis Projection persuit ICA.
Martina Uray Heinz Mayer Joanneum Research Graz Institute of Digital Image Processing Horst Bischof Graz University of Technology Institute for Computer.
- A Maximum Likelihood Approach Vinod Kumar Ramachandran ID:
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Experience Report: System Log Analysis for Anomaly Detection
Principal Component Analysis (PCA)
Semi-Supervised Clustering
Linli Xu Martha White Dale Schuurmans University of Alberta
School of Computer Science & Engineering
Optimum Passive Beamforming in Relation to Active-Passive Data Fusion
Segmentation of Dynamic Scenes from Image Intensities
Principal Component Analysis (PCA)
Machine Learning Basics
Unsupervised-learning Methods for Image Clustering
Outline Peter N. Belhumeur, Joao P. Hespanha, and David J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,”
PCA vs ICA vs LDA.
Pei Qi ECE at UW-Madison
Historic Document Image De-Noising using Principal Component Analysis (PCA) and Local Pixel Grouping (LPG) Han-Yang Tang1, Azah Kamilah Muda1, Yun-Huoy.
Outline S. C. Zhu, X. Liu, and Y. Wu, “Exploring Texture Ensembles by Efficient Markov Chain Monte Carlo”, IEEE Transactions On Pattern Analysis And Machine.
A survey of network anomaly detection techniques
Design of Hierarchical Classifiers for Efficient and Accurate Pattern Classification M N S S K Pavan Kumar Advisor : Dr. C. V. Jawahar.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Master Thesis Presentation
By Viput Subharngkasen
Presented by Nagesh Adluru
Unsupervised Learning II: Soft Clustering with Gaussian Mixture Models
EE513 Audio Signals and Systems
Dimension reduction : PCA and Clustering
Bootstrapping Jackknifing
INTRODUCTION TO Machine Learning
Feature space tansformation methods
Analytics – Statistical Approaches
Feature Selection Methods
Multivariate Methods Berlin Chen, 2005 References:
Improved Spread Spectrum: A New Modulation Technique for Robust Watermarking IEEE Trans. On Signal Processing, April 2003 Multimedia Security.
Julia Starke, Christian Eichmann, Simon Ottenhaus and Tamim Asfour
Spatial Signal Processing with Emphasis on Emitter Localization
Presentation transcript:

Outlier Processing via L1-Principal Subspaces Shubham Chamadia and Dimitris A. Pados Department of Electrical Engineering, The State University of New York at Buffalo, NY 14260 E-mail:{shubhamc, pados}@buffalo.edu

Outline Outlier detection classification Limitations of L1-PCA The proposed outlier processing scheme Simulation results Conclusion

What is an outlier? [Hawkins, 1980] an observation that “deviates so much from other observations as to arouse suspicions that it was generated by a different mechanism”. [Jarrell, 1994] a point that is far outside the norm for a variable or population. [Moore and McCabe, 1999] an observation that lies outside the overall pattern of distribution.

Outliers detection classification Supervised: Semi-supervised: Unsupervised: Training data with labeled normal and abnormal classes. Classification is highly unbalanced, labelling often done manually. Availability of labeled samples from only one class. Declare test instance not belonging to one class as belonging to other. Most preferred, makes no assumption about availability of labeled training samples. Assumption: normal instances are more frequent than abnormal. we focus on unsupervised technique

Unsupervised detection Principal Component Analysis (PCA) overview X2 Most valuable tool to reduce dimensionality of data. Finds new directions over which: Preserves most of samples’ information. Maximizes the data variance. X1 Problem formulation Given a real-valued data matrix Maximizes the -norm of projection of over

Unsupervised detection X2 Genesis of -PCA Outlier X1 Sensitive to extreme outliers. Nearest preferred solution often referred as - norm PCA (robust to outliers). Q. To what extent L1 –PCA is robust?

Motivation to outlier processing Setup Generate a 2D-data matrix Add 4 outlier points Limitation of and to some extent limitation of Urgent need for outlier removal in conjunction with robust

Proposed outlier processing via L1-PCs Step 1-4: For , obtain -principal components: Ways to calculate If sample support is small, use optimal -PCA algorithm[1] For medium to high sample, use iterative suboptimal algorithm[2]

Proposed outlier processing via L1-PCs Step 2-4: Evaluate reliability weight corresponding to each sample of X2 : often known as rank -reconstruction error of each sample Outlier : relative degree (normalized) of outliers Ideally, corrupted sample lie far away from -direction, high reconstruction error high weighted-value Þ X1

Proposed outlier processing via L1-PCs Step 3-4: Outlier detection and removal Prior knowledge of number of corrupted sample (say p) Discard p-highest weighted samples. Apriori knowledge often non-practical. No prior knowledge of corruption Implement (K=2)-means clustering over scalar reliability weight vector Extract sample index corresponding to higher mean cluster (potential outlier cluster) Discard samples contained in index set

Proposed outlier processing via L1-PCs Step 4-4: Recalculate -principal components over outlier processed data Expected reveals a deeper insight of given data matrix than

Proposed outlier processing via L1-PCs Step 1-4 Step 2-4 Scheme denoted by L1 + L1 Step 3-4 Step 4-4

Simulation result 1-3 Experiment 1: Data dimensionality reduction Setup Generate data matrix , i. i. d. Gaussian distribution Corrupt certain percentage of samples by outlier vector , i. i. d. Obtain P=2 principal components Metric: Average representation error (ARE) is the PC over corrupted data

Simulation result 2-3 Experiment 2: Direction-of-Arrival (DoA) estimation Setup Uniform linear antenna array D = 7 elements, recording N = 30 complex observation . A1 A7 A2 x : received signal amplitude, : Bernoulli equiprobable bits : additive white complex Gaussian noise vector : array response vector , with signal of interest at

Simulation result 2-3 Experiment 2: Direction-of-Arrival (DoA) estimation Setup cont. Adding jammer corruption . A1 A7 A2 x Number of corrupted samples = 3 (out of 30) Number of jammers (random location) corrupting = 3 Jammer SNR = 10 dB

Simulation result 2-3 Experiment 2: Direction-of-Arrival (DoA) estimation Setup cont. MUSIC-type DoA spectrum function: : principal component over corrupted data Jammer angle signal angle

Simulation result 2-3 Experiment 2: Direction-of-Arrival (DoA) estimation Setup cont. MUSIC-type DoA spectrum function: : principal component over corrupted data Jammer angle signal angle Root mean square error (RMSE) : estimated angle at realization

Simulation result 3-3 Experiment 3: Robust image fusion Setup Original image N = 10 grayscale image (256*256) Adding independent noise P10 Add AWGN, 𝜎 2 =50 P2 40% of 8 (out of 10) images with salt and pepper P1 P0 An extreme outlier Append with a baboon image

Simulation result 3-3 Experiment 3: Robust image fusion Processing . . Restoration block (32*32) P2 P1 P0 x x x Total restoration blocks 256 x 256 32 x 32 =64 x x . x . . . Proposed outlier scheme over x x x

Simulation result 3-3 Experiment 3: Robust image fusion original noisy Salt-pepper Baboon

Outlier processing via L1-PCs Conclusion: With advent of big-data, requires robust outlier detection technique. Conventional -subspaces have limitations to extreme outliers. Low complexity clustering over scalar reliability weights. Simulation exhibits the demand of such a robust outlier removal scheme.