Jure Leskovec and Christos Faloutsos Machine Learning Department

Slides:



Advertisements
Similar presentations
Eigen Decomposition and Singular Value Decomposition
Advertisements

Beyond Streams and Graphs: Dynamic Tensor Analysis
Eigen Decomposition and Singular Value Decomposition
Multilinear Algebra for Analyzing Data with Multiple Linkages
CMU SCS : Multimedia Databases and Data Mining Lecture #20: SVD - part III (more case studies) C. Faloutsos.
CMU SCS : Multimedia Databases and Data Mining Lecture #19: SVD - part II (case studies) C. Faloutsos.
Multilinear Algebra for Analyzing Data with Multiple Linkages Tamara G. Kolda plus: Brett Bader, Danny Dunlavy, Philip Kegelmeyer Sandia National Labs.
CMU SCS : Multimedia Databases and Data Mining Lecture #21: Tensor decompositions C. Faloutsos.
Dimensionality Reduction PCA -- SVD
Link Analysis: PageRank
15-826: Multimedia Databases and Data Mining
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 3 March 23, 2005
N EIGHBORHOOD F ORMATION AND A NOMALY D ETECTION IN B IPARTITE G RAPHS Jimeng Sun, Huiming Qu, Deepayan Chakrabarti & Christos Faloutsos Jimeng Sun, Huiming.
10-603/15-826A: Multimedia Databases and Data Mining SVD - part II (more case studies) C. Faloutsos.
Principal Component Analysis
Multimedia Databases SVD II. Optimality of SVD Def: The Frobenius norm of a n x m matrix M is (reminder) The rank of a matrix M is the number of independent.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 4 March 30, 2005
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 3 April 2, 2006
15-853Page :Algorithms in the Real World Indexing and Searching III (well actually II) – Link Analysis – Near duplicate removal.
Multimedia Databases SVD II. SVD - Detailed outline Motivation Definition - properties Interpretation Complexity Case studies SVD properties More case.
Link Analysis, PageRank and Search Engines on the Web
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
1 Algorithms for Large Data Sets Ziv Bar-Yossef Lecture 6 May 7, 2006
10-603/15-826A: Multimedia Databases and Data Mining SVD - part I (definitions) C. Faloutsos.
Link Structure and Web Mining Shuying Wang
Kathryn Linehan Advisor: Dr. Dianne O’Leary
Multimedia Databases LSI and SVD. Text - Detailed outline text problem full text scanning inversion signature files clustering information filtering and.
Singular Value Decomposition and Data Management
E.G.M. PetrakisDimensionality Reduction1  Given N vectors in n dims, find the k most important axes to project them  k is user defined (k < n)  Applications:
DATA MINING LECTURE 7 Dimensionality Reduction PCA – SVD
CMU SCS : Multimedia Databases and Data Mining Lecture #20: SVD - part III (more case studies) C. Faloutsos.
The PageRank Citation Ranking: Bringing Order to the Web Presented by Aishwarya Rengamannan Instructor: Dr. Gautam Das.
1 Information Retrieval through Various Approximate Matrix Decompositions Kathryn Linehan Advisor: Dr. Dianne O’Leary.
Introduction to tensor, tensor factorization and its applications
CS315 – Link Analysis Three generations of Search Engines Anchor text Link analysis for ranking Pagerank HITS.
Latent Semantic Indexing: A probabilistic Analysis Christos Papadimitriou Prabhakar Raghavan, Hisao Tamaki, Santosh Vempala.
SINGULAR VALUE DECOMPOSITION (SVD)
CMU SCS KDD '09Faloutsos, Miller, Tsourakakis P5-1 Large Graph Mining: Power Tools and a Practitioner’s guide Task 5: Graphs over time & tensors Faloutsos,
1 Latent Concepts and the Number Orthogonal Factors in Latent Semantic Analysis Georges Dupret
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
1 1 COMP5331: Knowledge Discovery and Data Mining Acknowledgement: Slides modified based on the slides provided by Lawrence Page, Sergey Brin, Rajeev Motwani.
Incremental Pattern Discovery on Streams, Graphs and Tensors Jimeng Sun Ph.D.Thesis Proposal May 15, 2006.
CMU SCS : Multimedia Databases and Data Mining Lecture #18: SVD - part I (definitions) C. Faloutsos.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
SCS CMU Speaker Hanghang Tong Colibri: Fast Mining of Large Static and Dynamic Graphs Speaking Skill Requirement.
Unsupervised Learning II Feature Extraction
Large Graph Mining: Power Tools and a Practitioner’s guide
CS479/679 Pattern Recognition Dr. George Bebis
15-826: Multimedia Databases and Data Mining
Zhu Han University of Houston Thanks for Dr. Hung Nguyen’s Slides
15-499:Algorithms and Applications
Search Engines and Link Analysis on the Web
15-826: Multimedia Databases and Data Mining
LSI, SVD and Data Management
Packing to fewer dimensions
Part 2: Graph Mining Tools - SVD and ranking
Lecture 22 SVD, Eigenvector, and Web Search
Prof. Paolo Ferragina, Algoritmi per "Information Retrieval"
SVD: Physical Interpretation and Applications
Dimensionality Reduction: SVD & CUR
Jimeng Sun · Charalampos (Babis) E
Prof. Paolo Ferragina, Algoritmi per "Information Retrieval"
Roadmap Introduction – Motivation Part#1: Graphs
15-826: Multimedia Databases and Data Mining
Next Generation Data Mining Tools: SVD and Fractals
Junghoo “John” Cho UCLA
Lecture 22 SVD, Eigenvector, and Web Search
Packing to fewer dimensions
Lecture 22 SVD, Eigenvector, and Web Search
Lecture 20 SVD and Its Applications
Presentation transcript:

Jure Leskovec and Christos Faloutsos Machine Learning Department Tools for large graph mining WWW 2008 tutorial Part 3: Matrix tools for graph mining Jure Leskovec and Christos Faloutsos Machine Learning Department Joint work with: Deepay Chakrabarti, Tamara Kolda and Jimeng Sun.

Leskovec&Faloutsos, WWW 2008 Tutorial outline Part 1: Structure and models for networks What are properties of large graphs? How do we model them? Part 2: Dynamics of networks Diffusion and cascading behavior How do viruses and information propagate? Part 3: Matrix tools for mining graphs Singular value decomposition (SVD) Random walks Part 4: Case studies 240 million MSN instant messenger network Graph projections: how does the web look like Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun About part 3 Introduce matrix and tensor tools through real mining applications Goal: find patterns, rules, clusters, outliers, … in matrices and in tensors Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 What is this part about? Connection of matrix tools and networks Matrix tools Singular Value Decomposition (SVD) Principal Component Analysis (PCA) Webpage ranking algorithms: HITS, PageRank CUR decomposition Co-clustering Tensor tools Tucker decomposition Applications Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Why matrices? Examples Social networks Documents and terms Authors and terms Peter Mary Nick John ... John Peter Mary Nick ... 11 22 55 ... 5 6 7 Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Why matrices? Examples Social networks Documents and terms Authors and terms Peter Mary Nick John ... John Peter Mary Nick ... 11 22 55 ... 5 6 7 Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Why tensors? Example Tensor: n-dimensional generalization of matrix SIGMOD’07 data mining classif. tree ... John Peter Mary Nick Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Why tensors? Example Tensor: n-dimensional generalization of matrix SIGMOD’05 SIGMOD’06 SIGMOD’07 data mining classif. tree ... John Peter Mary Nick Leskovec&Faloutsos, WWW 2008

Tensors are useful for 3 or more modes Faloutsos, Kolda, Sun Tensors are useful for 3 or more modes Terminology: ‘mode’ (or ‘aspect’): Mode#3 data mining classif. tree ... Mode#2 Mode (== aspect) #1 Leskovec&Faloutsos, WWW 2008

Motivating applications Faloutsos, Kolda, Sun Motivating applications Social networks Why matrices are important? Why tensors are useful? P1: social networks P2: web & text mining P3: network forensics P4: sensor networks Sensor networks Network forensics Temperature normal traffic abnormal traffic destination source Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun Static Data model Tensor Formally, Generalization of matrices Represented as multi-array, (~ data cube). Order 1st 2nd 3rd Correspondence Vector Matrix 3D array Example Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun Dynamic Data model Tensor Streams A sequence of Mth order tensors where t is increasing over time Order 1st 2nd 3rd Correspondence Multiple streams Time evolving graphs 3D arrays Example keyword time … time author … … Leskovec&Faloutsos, WWW 2008

SVD: Examples of Matrices Faloutsos, Kolda, Sun SVD: Examples of Matrices Example/Intuition: Documents and terms Find patterns, groups, concepts data mining classif. tree ... Paper#1 Paper#2 Paper#3 Paper#4 ... Leskovec&Faloutsos, WWW 2008

Singular Value Decomposition (SVD) Faloutsos, Kolda, Sun Singular Value Decomposition (SVD) X = UVT X U x(1) x(2) x(M) u1 u2 uk  VT v1 v2 vk 1 2 . . = k right singular vectors singular values input data left singular vectors Leskovec&Faloutsos, WWW 2008

SVD as spectral decomposition Faloutsos, Kolda, Sun SVD as spectral decomposition Best rank-k approximation in L2 and Frobenius SVD only works for static matrices (a single 2nd order tensor) n n 1u1v1 2u2v2 m A   VT + m U Leskovec&Faloutsos, WWW 2008

Vector outer product – intuition: owner age 20; 30; 40 20; 30; 40 car type VW Volvo BMW VW Volvo BMW A 1-d histograms + independence assumption 2-d histogram Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Example A = U  VT - example: retrieval inf. lung brain data CS x x = MD Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Example A = U  VT - example: retrieval CS-concept inf. lung MD-concept brain data CS x x = MD Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Example doc-to-concept similarity matrix A = U  VT - example: retrieval CS-concept inf. lung MD-concept brain data CS x x = MD Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Example A = U  VT - example: retrieval ‘strength’ of CS-concept inf. lung brain data CS x x = MD Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Example A = U  VT - example: term-to-concept similarity matrix retrieval inf. lung brain data CS-concept CS x x = MD Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Example A = U  VT - example: term-to-concept similarity matrix retrieval inf. lung brain data CS-concept CS x x = MD Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD - Interpretation ‘documents’, ‘terms’ and ‘concepts’: Q: if A is the document-to-term matrix, what is AT A? A: term-to-term ([m x m]) similarity matrix Q: A AT ? A: document-to-document ([n x n]) similarity matrix Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun SVD properties V are the eigenvectors of the covariance matrix ATA U are the eigenvectors of the Gram (inner-product) matrix AAT Further reading: 1. Ian T. Jolliffe, Principal Component Analysis (2nd ed), Springer, 2002. 2. Gilbert Strang, Linear Algebra and Its Applications (4th ed), Brooks Cole, 2005. Leskovec&Faloutsos, WWW 2008

Principal Component Analysis (PCA) Faloutsos, Kolda, Sun Principal Component Analysis (PCA) SVD PCA is an important application of SVD Note that U and V are dense and may have negative entries n n R k k k VT A U  m m Loading PCs Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun PCA interpretation best axis to project on: (‘best’ = min sum of squares of projection errors) Term2 (‘lung’) Term1 (‘data’) Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun PCA - interpretation  U VT Term2 (‘retrieval’) PCA projects points Onto the “best” axis first singular vector v1 minimum RMS error Term1 (‘data’) Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm HITS Problem definition: given the web and a query find the most ‘authoritative’ web pages for this query Step 0: find all pages containing the query terms Step 1: expand by one move forward and backward Further reading: 1. J. Kleinberg. Authoritative sources in a hyperlinked environment. SODA 1998 Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm HITS Step 1: expand by one move forward and backward Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm HITS on the resulting graph, give high score (= ‘authorities’) to nodes that many important nodes point to give high importance score (‘hubs’) to nodes that point to good ‘authorities’ hubs authorities Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm HITS observations recursive definition! each node (say, ‘i’-th node) has both an authoritativeness score ai and a hubness score hi Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm: HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm: HITS Let A be the adjacency matrix: the (i,j) entry is 1 if the edge from i to j exists Let h and a be [n x 1] vectors with the ‘hubness’ and ‘authoritativiness’ scores. Then: Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm: HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm: HITS Then: ai = hk + hl + hm that is ai = Sum (hj) over all j that (j,i) edge exists or a = AT h k i l m Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm: HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm: HITS symmetrically, for the ‘hubness’: hi = an + ap + aq that is hi = Sum (qj) over all j that (i,j) edge exists or h = A a i n p q Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm: HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm: HITS In conclusion, we want vectors h and a such that: h = A a a = AT h That is: a = ATA a Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm: HITS Faloutsos, Kolda, Sun Kleinberg’s algorithm: HITS a is a right singular vector of the adjacency matrix A (by dfn!), a.k.a the eigenvector of ATA Starting from random a’ and iterating, we’ll eventually converge Q: to which of all the eigenvectors? why? A: to the one of the strongest eigenvalue, (AT A ) k a = 1ka Leskovec&Faloutsos, WWW 2008

Kleinberg’s algorithm - discussion Faloutsos, Kolda, Sun Kleinberg’s algorithm - discussion ‘authority’ score can be used to find ‘similar pages’ (how?) closely related to ‘citation analysis’, social networks / ‘small world’ phenomena Leskovec&Faloutsos, WWW 2008

Motivating problem: PageRank Faloutsos, Kolda, Sun Motivating problem: PageRank Given a directed graph, find its most interesting/central node A node is important, if it is connected with important nodes (recursive, but OK!) Leskovec&Faloutsos, WWW 2008

Motivating problem – PageRank solution Faloutsos, Kolda, Sun Motivating problem – PageRank solution Given a directed graph, find its most interesting/central node Proposed solution: Random walk; spot most ‘popular’ node (-> steady state prob. (ssp)) A node has high ssp, if it is connected with high ssp nodes (recursive, but OK!) Leskovec&Faloutsos, WWW 2008

(Simplified) PageRank algorithm Let A be the transition matrix (= adjacency matrix); let B be the transpose, column-normalized - then From B To 1 2 3 4 5 = Leskovec&Faloutsos, WWW 2008

(Simplified) PageRank algorithm B p = p B p = p 2 1 3 = 4 5 Leskovec&Faloutsos, WWW 2008

(Simplified) PageRank algorithm Faloutsos, Kolda, Sun (Simplified) PageRank algorithm B p = 1 * p thus, p is the eigenvector that corresponds to the highest eigenvalue (=1, since the matrix is column-normalized) Why does such a p exist? p exists if B is nxn, nonnegative, irreducible [Perron–Frobenius theorem] Leskovec&Faloutsos, WWW 2008

(Simplified) PageRank algorithm Faloutsos, Kolda, Sun (Simplified) PageRank algorithm In short: imagine a particle randomly moving along the edges compute its steady-state probabilities (ssp) Full version of algo: with occasional random jumps Why? To make the matrix irreducible Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun Full Algorithm With probability 1-c, fly-out to a random node Then, we have p = c B p + (1-c)/n 1 => p = (1-c)/n [I - c B] -1 1 Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008

Motivation of CUR or CMD Faloutsos, Kolda, Sun Motivation of CUR or CMD SVD, PCA all transform data into some abstract space (specified by a set basis) Interpretability problem Loss of sparsity Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun PCA - interpretation Term2 (‘retrieval’) PCA projects points Onto the “best” axis first singular vector v1 minimum RMS error Term1 (‘data’) Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun CUR Example-based projection: use actual rows and columns to specify the subspace Given a matrix ARmn, find three matrices C Rmc, U Rcr, R Rr n , such that ||A-CUR|| is small Orthogonal projection U is the pseudo-inverse of X Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun CUR Example-based projection: use actual rows and columns to specify the subspace Given a matrix ARmn, find three matrices C Rmc, U Rcr, R Rr n , such that ||A-CUR|| is small Example-based U is the pseudo-inverse of X: U = X† = (UT U )-1 UT Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun CUR (cont.) Key question: How to select/sample the columns and rows? Uniform sampling Biased sampling CUR w/ absolute error bound CUR w/ relative error bound Reference: Tutorial: Randomized Algorithms for Matrices and Massive Datasets, SDM’06 Drineas et al. Subspace Sampling and Relative-error Matrix Approximation: Column-Row-Based Methods, ESA2006 Drineas et al., Fast Monte Carlo Algorithms for Matrices III: Computing a Compressed Approximate Matrix Decomposition, SIAM Journal on Computing, 2006. Leskovec&Faloutsos, WWW 2008

The sparsity property – pictorially: Faloutsos, Kolda, Sun The sparsity property – pictorially: SVD/PCA: Destroys sparsity = U S VT CUR: maintains sparsity animation = C U R Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun The sparsity property sparse and small SVD: A = U  VT Big but sparse Big and dense dense but small CUR: A = C U R Big but sparse Big but sparse Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Matrix tools - summary SVD: optimal for L2 – VERY popular (HITS, PageRank, Karhunen-Loeve, Latent Semantic Indexing, PCA, etc etc) C-U-R (CMD etc) near-optimal; sparsity; interpretability Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 TENSORS Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun Reminder: SVD Best rank-k approximation in L2 n n m A   VT m U Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun Reminder: SVD Best rank-k approximation in L2 n 1u1v1 2u2v2 A  + m Leskovec&Faloutsos, WWW 2008

Goal: extension to >=3 modes Faloutsos, Kolda, Sun Goal: extension to >=3 modes I x R K x R A B J x R C R x R x R I x J x K +…+ = ¼ Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Main points: 2 major types of tensor decompositions: PARAFAC and Tucker both can be solved with ``alternating least squares’’ (ALS) Details follow – we start with terminology: Leskovec&Faloutsos, WWW 2008

A tensor is a multidimensional array Faloutsos, Kolda, Sun A tensor is a multidimensional array Column (Mode-1) Fibers Row (Mode-2) Fibers Tube (Mode-3) Fibers An I £ J £ K tensor K xijk I J Horizontal Slices Lateral Slices Frontal Slices 3rd order tensor mode 1 has dimension I mode 2 has dimension J mode 3 has dimension K Before we can get into multilinear algebra and its applications, we need to talk about multilinear arrays. Note: All my examples will be in 3d, but it all extends to higher dimensions as well. Note: Tutorial focus is on 3rd order, but everything can be extended to higher orders. Leskovec&Faloutsos, WWW 2008

Tucker Decomposition - intuition Faloutsos, Kolda, Sun Tucker Decomposition - intuition I x J x K ¼ A I x R B J x S C K x T R x S x T author x keyword x conference A: author x author-group B: keyword x keyword-group C: conf. x conf-group G: how groups relate to each other Leskovec&Faloutsos, WWW 2008

Matrization: Converting a Tensor to a Matrix Faloutsos, Kolda, Sun Matrization: Converting a Tensor to a Matrix (i,j,k) (i′,j′) Matricize (unfolding) Reverse Matricize X(n): The mode-n fibers are rearranged to be the columns of a matrix 5 7 6 8 1 3 2 4 Vectorization Leskovec&Faloutsos, WWW 2008

Tensor Mode-n Multiplication Faloutsos, Kolda, Sun Tensor Mode-n Multiplication Tensor Times Matrix Tensor Times Vector Compute the dot product of a and each column (mode-1) fiber Multiply each row (mode-2) fiber by B Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Tensor tools - summary Two main tools PARAFAC Tucker Both find row-, column-, tube-groups but in PARAFAC the three groups are identical To solve: Alternating Least Squares Toolbox: from Tamara Kolda: http://csmr.ca.sandia.gov/~tgkolda/TensorToolbox/ Leskovec&Faloutsos, WWW 2008

P1: Environmental sensor monitoring Faloutsos, Kolda, Sun P1: Environmental sensor monitoring Light Temperature Voltage Humidity Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun P1: sensor monitoring 1st factor Scaling factor 250 time location type Location Type Time light voltage 1st factor consists of the main trends: Daily periodicity on time Uniform on all locations Temp, Light and Volt are positively correlated while negatively correlated with Humid hum. temp. Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun P1: sensor monitoring type time location 2nd factor Scaling factor 154 2nd factor captures an atypical trend: Uniformly across all time Concentrating on 3 locations Mainly due to voltage Interpretation: two sensors have low battery, and the other one has high battery. light voltage hum. temp. Leskovec&Faloutsos, WWW 2008

P3: Social network analysis Faloutsos, Kolda, Sun P3: Social network analysis Multiway latent semantic indexing (LSI) Monitor the change of the community structure over time Philip Yu time Michael Stonebreaker ‘Pattern’ ‘Query’ Leskovec&Faloutsos, WWW 2008

P3: Social network analysis (cont.) Faloutsos, Kolda, Sun P3: Social network analysis (cont.) Authors Keywords Year michael carey, michael stonebreaker, h. jagadish, hector garcia-molina queri,parallel,optimization,concurr, objectorient 1995 surajit chaudhuri,mitch cherniack,michael stonebreaker,ugur etintemel distribut,systems,view,storage,servic,process,cache 2004 jiawei han,jian pei,philip s. yu, jianyong wang,charu c. aggarwal streams,pattern,support, cluster, index,gener,queri DB DM Two groups are correctly identified: Databases and Data mining People and concepts are drifting over time Leskovec&Faloutsos, WWW 2008

P4: Network anomaly detection Faloutsos, Kolda, Sun P4: Network anomaly detection Abnormal traffic Reconstruction error over time Normal traffic Reconstruction error gives indication of anomalies. Prominent difference between normal and abnormal ones is mainly due to the unusual scanning activity (confirmed by the campus admin). Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun P5: Web graph mining How to order the importance of web pages? Kleinberg’s algorithm HITS PageRank Tensor extension on HITS (TOPHITS) 8 mins Leskovec&Faloutsos, WWW 2008

Kleinberg’s Hubs and Authorities (the HITS method) Faloutsos, Kolda, Sun Kleinberg’s Hubs and Authorities (the HITS method) Sparse adjacency matrix and its SVD: authority scores for 2nd topic authority scores for 1st topic to from hub scores for 1st topic hub scores for 2nd topic Kleinberg, JACM, 1999 Leskovec&Faloutsos, WWW 2008

HITS Authorities on Sample Data Faloutsos, Kolda, Sun HITS Authorities on Sample Data We started our crawl from http://www-neos.mcs.anl.gov/neos, and crawled 4700 pages, resulting in 560 cross-linked hosts. authority scores for 2nd topic authority scores for 1st topic to from hub scores for 1st topic hub scores for 2nd topic Leskovec&Faloutsos, WWW 2008

Three-Dimensional View of the Web Faloutsos, Kolda, Sun Three-Dimensional View of the Web Observe that this tensor is very sparse! Kolda, Bader, Kenny, ICDM05 Leskovec&Faloutsos, WWW 2008

Topical HITS (TOPHITS) Faloutsos, Kolda, Sun Topical HITS (TOPHITS) Main Idea: Extend the idea behind the HITS model to incorporate term (i.e., topical) information. term scores for 1st topic term scores for 2nd topic term to from authority scores for 1st topic authority scores for 2nd topic hub scores for 2nd topic hub scores for 1st topic Leskovec&Faloutsos, WWW 2008

Topical HITS (TOPHITS) Faloutsos, Kolda, Sun Topical HITS (TOPHITS) Main Idea: Extend the idea behind the HITS model to incorporate term (i.e., topical) information. term scores for 1st topic term scores for 2nd topic term to from authority scores for 1st topic authority scores for 2nd topic hub scores for 2nd topic hub scores for 1st topic Leskovec&Faloutsos, WWW 2008

TOPHITS Terms & Authorities on Sample Data Faloutsos, Kolda, Sun TOPHITS Terms & Authorities on Sample Data TOPHITS uses 3D analysis to find the dominant groupings of web pages and terms. wk = # unique links using term k Tensor term scores for 1st topic term scores for 2nd topic term to from authority scores for 1st topic authority scores for 2nd topic hub scores for 2nd topic hub scores for 1st topic Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 Faloutsos, Kolda, Sun Conclusions Real data are often in high dimensions with multiple aspects (modes) Matrices and tensors provide elegant theory and algorithms Several research problems are still open skewed distribution, anomaly detection, streaming algorithms, distributed/parallel algorithms, efficient out-of-core processing Leskovec&Faloutsos, WWW 2008

Leskovec&Faloutsos, WWW 2008 References Slides borrowed from SIGMOD ‘07 tutorial by Falutsos, Kolda and Sun. Leskovec&Faloutsos, WWW 2008