Additive Data Perturbation: data reconstruction attacks.

Slides:



Advertisements
Similar presentations
Covariance Matrix Applications
Advertisements

Tensors and Component Analysis Musawir Ali. Tensor: Generalization of an n-dimensional array Vector: order-1 tensor Matrix: order-2 tensor Order-3 tensor.
1er. Escuela Red ProTIC - Tandil, de Abril, 2006 Principal component analysis (PCA) is a technique that is useful for the compression and classification.
Principal Components Analysis Babak Rasolzadeh Tuesday, 5th December 2006.
1cs542g-term High Dimensional Data  So far we’ve considered scalar data values f i (or interpolated/approximated each component of vector values.
Principal Component Analysis CMPUT 466/551 Nilanjan Ray.
Principal Component Analysis
Eigenvalues and eigenvectors
Principal component analysis (PCA)
Dimensional reduction, PCA
SAC’06 April 23-27, 2006, Dijon, France On the Use of Spectral Filtering for Privacy Preserving Data Mining Songtao Guo UNC Charlotte Xintao Wu UNC Charlotte.
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 9(b) Principal Components Analysis Martin Russell.
Face Recognition Jeremy Wyatt.
Independent Component Analysis (ICA) and Factor Analysis (FA)
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
1 Deriving Private Information from Randomized Data Zhengli Huang Wenliang (Kevin) Du Biao Chen Syracuse University.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Ordinary least squares regression (OLS)
1 When Does Randomization Fail to Protect Privacy? Wenliang (Kevin) Du Department of EECS, Syracuse University.
Principal Component Analysis Principles and Application.
The Multivariate Normal Distribution, Part 2 BMTRY 726 1/14/2014.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Summarized by Soo-Jin Kim
Principle Component Analysis Presented by: Sabbir Ahmed Roll: FH-227.
Linear Least Squares Approximation. 2 Definition (point set case) Given a point set x 1, x 2, …, x n  R d, linear least squares fitting amounts to find.
Chapter 2 Dimensionality Reduction. Linear Methods
Presented By Wanchen Lu 2/25/2013
Principles of Pattern Recognition
Principal Component Analysis Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Multivariate Statistics Matrix Algebra I W. M. van der Veld University of Amsterdam.
N– variate Gaussian. Some important characteristics: 1)The pdf of n jointly Gaussian R.V.’s is completely described by means, variances and covariances.
A note about gradient descent: Consider the function f(x)=(x-x 0 ) 2 Its derivative is: By gradient descent (If f(x) is more complex we usually cannot.
Principal Component Analysis (PCA). Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite)
ECE 8443 – Pattern Recognition LECTURE 08: DIMENSIONALITY, PRINCIPAL COMPONENTS ANALYSIS Objectives: Data Considerations Computational Complexity Overfitting.
NCAF Manchester July 2000 Graham Hesketh Information Engineering Group Rolls-Royce Strategic Research Centre.
Neural Computation Prof. Nathan Intrator
Matrix Notation for Representing Vectors
EIGENSYSTEMS, SVD, PCA Big Data Seminar, Dedi Gadot, December 14 th, 2014.
Principle Component Analysis and its use in MA clustering Lecture 12.
Principal Component Analysis (PCA)
Presented by: Muhammad Wasif Laeeq (BSIT07-1) Muhammad Aatif Aneeq (BSIT07-15) Shah Rukh (BSIT07-22) Mudasir Abbas (BSIT07-34) Ahmad Mushtaq (BSIT07-45)
Principal Components Analysis ( PCA)
Unsupervised Learning II Feature Extraction
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
1 C.A.L. Bailer-Jones. Machine Learning. Data exploration and dimensionality reduction Machine learning, pattern recognition and statistical data modelling.
Introduction to Vectors and Matrices
Information Management course
Matrices and vector spaces
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
LECTURE 10: DISCRIMINANT ANALYSIS
9.3 Filtered delay embeddings
CH 5: Multivariate Methods
Principal Component Analysis (PCA)
Principal Component Analysis
Additive Data Perturbation: data reconstruction attacks
Dynamic graphics, Principal Component Analysis
The regression model in matrix form
Techniques for studying correlation and covariance structure
Recitation: SVD and dimensionality reduction
Parallelization of Sparse Coding & Dictionary Learning
Feature space tansformation methods
Principal Components What matters most?.
Digital Image Processing Lecture 21: Principal Components for Description Prof. Charlene Tsai *Chapter 11.4 of Gonzalez.
LECTURE 09: DISCRIMINANT ANALYSIS
Introduction to Vectors and Matrices
Feature Selection Methods
Principal Component Analysis
Lecture 16. Classification (II): Practical Considerations
Marios Mattheakis and Pavlos Protopapas
Outline Variance Matrix of Stochastic Variables and Orthogonal Transforms Principle Component Analysis Generalized Eigenvalue Decomposition.
Presentation transcript:

Additive Data Perturbation: data reconstruction attacks

Outline  Overview Paper “Deriving Private Information from Randomized Data”  Data Reconstruction Methods PCA-based method Bayes method  Comparison  Summary

Overview  Data reconstruction Z = X+R Problem: know Z and distribution of R  estimate the value of X Extend it to matrix  X contains multiple dimensions  Or folding the vector X  matrix

Two major approaches  Principle component analysis (PCA) based approach  Bayes analysis approach

Variance and covariance  Definition Random variable x, mean  Var(x) = E[(x- ) 2 ] Cov(xi, xj) = E[(xi- i)(xj- j)] For multidimensional case,  X=(x1,x2,…,xm)  Covariance matrix  If each dimension xi has mean zero cov(X) = 1/n X T *X

PCA intuition  Vector in space Original space  base vectors E={e1,e2,…,em}  Example: 3-dimension space x,y,z axes corresponds to {(1 0 0),(0 1 0), (0 0 1)} If we want to use the red axes to represent the vectors  The new base vectors U=(u1, u2)  Transformation: matrix X  XU X1X1 X2X2 u1 u2

 Why do we want to use different bases? Actual data distribution can be possibly described with lower dimensions X1X1 X2X2 u1 Ex: projecting points to U1, we can use one dimension (u1) to approximately describe all these points  The key problem: finding these directions that maximize variance of the points. These directions are called principle components.

How to do PCA?  Calculating covariance matrix: C =  “Eigenvalue decomposition” on C Matrix C: symmetric We can always find an orthonormal matrix U  U*U T = I  So that C = U*B*U T  B is a diagonal matrix X is normalized to mean zero for each dimension; n is the number of rows in X

Explanation of PCA  Explanation: di in B are actually the variance in the transformed space, and U is the transformation matrix 1/n X T *X =U*B*U T  1/n (XU) T *(XU) =B

 Look at the diagonal matrix B (eigenvalues) We know the variance in each transformed direction We can select the maximum ones (e.g., k of d elements) to approximately describe the total variance  Approximation with maximum eigenvalues Select the corresponding k eigenvectors in U  U’ Transform X  XU’  XU’ has only k dimensional  Use of PCA Dimensionality reduction Noise filtering

PCA-based reconstruction  Cov matrix for Y=X+R Elements in R is iid with variance  2 Cov(Xi+Ri, Xj+Rj) = cov(Xi,Xi) +  2, for diagonal elements cov(Xi,Xj) for i!=j Therefore, removing  2 from the diagonal of cov(Y), we get the covariance matrix for X

 Reconstruct X We have got C=cov(X) Apply PCA on cov matrix C  C = U*B*U T Select major principle components and get the corresponding eigenvectors U’ Reconstruct X  X^ = Y*U’*U’ T  Understanding it: X’: X in transformed space X’ =X*U  X=X’*U -1 =X’*U T ~ X’*U’ T approximate X’ with Y*U’ and plugin Error comes from here

Error analysis  X^ = Y*U’*U’ T  X^ = (X+R)*U’*U’ T  The error item is R*U’*U’ T  Mean square error is used to evaluate the quality of estimation xi and xi^ is single data item and its estimation: MSE = sum (xi-xi^) 2  Result: MSE = p/m *  2, is the variance of the noise

Bayes Method  Make an assumption The original data is multidimensional normal distribution The noise is is also normal distribution Covariance matrix, can be approximated with the discussed method.

 Data (x11,x12,…x1m)  vector (x21,x22,…x2m)  vector …

 Problem: Given a vector yi, yi=xi+ri Find the vector xi Maximize the posterior prob P(X|Y)

 Again, applying bayes rule f Constant for all x Maximize this With f y|x (y|x) = f R (y-x), plug in the distributions fx and f R We find x to maximize:

 It’s equivalent to maximize the exponential part  A function is maximized/minimized, when its derivative =0 i.e., Solving the above equation, we get

 Reconstruction For each vector y, plug in the covariance, the mean of vector x, and the noise variance, we get the estimate of the corresponding x

Experiments  Errors vs. number of dimensions Conclusion: covariance between dimensions helps reduce errors

 Errors vs. # of principle components # of PC : the correlation between dimensions Conclusion: the # of principal components ~ the amount of noise

Discussion  The key: find the covariance matrix of the original data X Increase the difficulty of Cov(X) estimation  decrease the accuracy of data reconstruction  Assumption of normal distribution for the Bayes method other distributions?