Download presentation

Presentation is loading. Please wait.

Published byCalvin Danis Modified about 1 year ago

1
Fast Jensen-Shannon Graph Kernel Bai Lu and Edwin Hancock Department of Computer Science University of York Supported by a Royal Society Wolfson Research Merit Award

2
Structural Variations

3
Protein-Protein Interaction Networks

4
Manipulating graphs Is structure similar (graph isomorphism, inexact match)? Is complexity similar (are graphs from same class but different in detail)? Is complexity (type of structure) uniform?

5
Goals Can we capture determine the similarity of structure using measures that capture their intrinsic complexity. Can graph entropies be used for this purpose. If they can then they lead naturally to information theoretic kernels and description length for learning over graph data.

6
Outline Literature Review: State of the Art Graph Kernels Existing graph kernel methods ： Graph kernels based on a) walks, b) paths or c) subgraph or subtree structures. Prior Work: Recently we have developed on information theoretic graph kernel based on Jensen-Shannon divergence probability distributions on graphs. Fast Jensen-Shannon Graph Kernel: Based on depth depth-based subgraph representation of a graph Based around graph centroid Experiments Conclusion

7
Literature Review: Graph Kernels Existing Graph Kernels (i.e Graph Kernels from the R- convolution [Haussler, 1999]) fall into three classes: Restricted subgraph or subtree kernels Weisfeiler-Lehman subtree kernel [Shevashidze et al., 2009, NIPS] Random walk kernels Product graph kernels [Gartner et al., 2003, ICML] Marginalized kernels on graphs [Kashima et al., 2003, ICML] Path based kernels Shortest path kernel [Borgwardt, 2005, ICDM]

8
Motivation Limitations of existing graph kernel Can not scale up to substructures of large size (e.g. (sub)graphs with hundreds or even thousands vertices). Compromised to substructures of limited size and only roughly capture topological arrangement within a graph. Even for relatively small subgraphs, most graph kernels still require significant computational overheads. Aim: develop a novel subgraph kernel for efficient computation, even when a pair of fully sized subgraphs are compared.

9
Approach Investigate how to kernelize depth-based graph representations by similarity for K-layer subgraphs using the Jensen-Shannon divergence. Commence by showing how to compute a fast Jensen- Shannon diffusion kernel for a pair of (sub)graphs. Describe how to compute a fast depth-based graph representation., based on complexity of structure. Combine ideas to compute fast Jensen-Shannon subgraph kernel.

10
Notation Consider a graph, adjacency matrix has elements The vertex degree matrix of is given by Normalaised Laplacian and its spectrum

11
The Jensen-Shannon Diffusion Kernel Jensen-Shannon diffusion kernel for graphs: For graphs Gp and Gq, the Jensen-Shannon divergence is where is entropy of composite structure formed from two (sub)graphs being compared (here we use the disjoint union). The Jensen-Shannon diffusion kernel for Gp and Gq is where entropy H(·) is either Shannon or the von Neumann.

12
Composite Structure Composite entropy of disjoint union A disjoint union of a pair of graph of graphs G p and G q is Graphs G p and G q are the connected components of the disjoint union graph G DU. Let p = |V p |/|V DU | and q = |V q |/|V DU |. Entropy (i.e. the composite entropy) of G DU is

13
Graph Entropy: Measures of complexity Shannon entropy of random walk : The probability of a steady state random walk on visiting vertex v i is. Shannon entropy of steady state random walk is von Neumann entropy: entropy associated with normalised Laplacian eigenvalues. Approximated by (Han PRL12)

14
Properties The Jensen-Shannon diffusion kernel for graphs : The Jensen-Shannon diffusion kernel is positive definite (pd). This follows the definitions in [Kondor and Lafferty, 2002, ICML], if a dissimilarity measure between a pair of graphs Gp and Gq satisfies symmetry, then a diffusion kernel associated with the similarity measure is pd. Time Complexity: For a pair of graphs Gp and Gq both having n vertices, computing the Jensen-Shannon diffusion kernel requires time complexity O(n^2).

15
Idea Decompose graph into layered subgraphs from centroid. Use JSD to compare subgraphs. Construct kernel over subgraphs.

16
The Depth-Based Representation of A Graph Subgraphs from the Centroid Vertex For graph G(V,E), construct shortest path matrix matrix SG whose element SG(i, j) are the shortest path lengths between vertices vi and vj. Average-shortest-path vector SV for G(V,E) is a vector with element from vertex vi to the remaining vertices. Centroid vertex for G(V,E) as The K-layer centroid expansion subgraph where

17
Depth-Based Representation For a graph G, we obtain a family of centroid expansion subgraphs, the depth- based representation of G is defined as where H(·) is either the Shannon entropy or the von Neumann entropy. Measures complexity via variation of entropy with depth

18
The Depth-Based Representation An example of the depth-based representation for a graph from the centroid vertex

19
Fast Jensen-Shannon Subgraph Kernel For a pair of graphs Gp(Vp, Ep) and Gq(Vq, Eq), similarity measure is is summed over an entropy-based similarity measure for the K-layer subgraphs. Jensen-Shannon diffusion kernel is the sum of the diffusion kernel measures for all the pairs of K-layer subgraphs Jensen-Shannon subgraph kernel is pd. Because, the proposed subgraph kernel is the sum of the positive Jensen-Shannon diffusion kernel.

20
Times Complexity Subgraph kernel graphs for graphs with n vertices and m edges, has time complexity O(n^2L + mn), where L is the size of the largest layer of the expansion subgraph. Depth–based representation is O(n^2L+mn). Jensen-Shannon diffusion kernel is O(n^2).

21
Observations Advantages a) von Neumann entropy is associated with the degree variance of connected vertices. Subgraph kernel is sensitive to interconnections between vertex clusters. b) For Shannon entropy vertices with large degrees dominate the entropy. Subgraph kernel is suited to characterizing a group of highly interconnected vertices, i.e. a dominant cluster. c) The depth-based representation captures inhomogeneities of complexity with depth. Enables it go gauge structure more finely than straightforwardly applying Jensen-Shannon diffusion kernel to original graphs. d) The proposed subgraph kernel only compares the pairs of subgraphs with the same layer size K. Avoids enumerating all the pairs of subgraphs and renders an efficient computation. e) Overcomes the subgraph size restriction which arises in existing graph kernels.

22
Experiments ( New, not in the paper ) We evaluate the classification performance of our kernel using 10-fold cross validation associated with C-Support Vector Machine. (Intel i5 3210M 2.5GHz) Classification of graphs abstracted from bioinformatics and computer vision databases. This datasets include: GatorBait (3D shapes), DD, COIL5 (images), CATH1, CATH2. Graph kernels for comparisons include: a) our kernel: 1) using the Shannon entropy (JSSS) 2) using the von Neumann entropy (JSSV) b) Weisfeiler-Lehman subtree kernel (WL), c) the shortest path graph kernel (SPGK), d) the graphlet count kernel (GCGK)

23
Experiments Details of the datasets

24
Experiments Classification Timing

25
Conclusion and Further Work Conclusion Presented a fast version of our Jensen-Shannon kernel. Compares well to alternatives on standard ML datasets. Further Work Hypergraphs, alternative entropies and divergences.

Similar presentations

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google