Kijung Shin Jinhong Jung Lee Sael U Kang

Slides:



Advertisements
Similar presentations
Weiren Yu 1, Jiajin Le 2, Xuemin Lin 1, Wenjie Zhang 1 On the Efficiency of Estimating Penetrating Rank on Large Graphs 1 University of New South Wales.
Advertisements

BiG-Align: Fast Bipartite Graph Alignment
+ Multi-label Classification using Adaptive Neighborhoods Tanwistha Saha, Huzefa Rangwala and Carlotta Domeniconi Department of Computer Science George.
Fast Algorithms For Hierarchical Range Histogram Constructions
TI: An Efficient Indexing Mechanism for Real-Time Search on Tweets Chun Chen 1, Feng Li 2, Beng Chin Ooi 2, and Sai Wu 2 1 Zhejiang University, 2 National.
Yasuhiro Fujiwara (NTT Cyber Space Labs)
The math behind PageRank A detailed analysis of the mathematical aspects of PageRank Computational Mathematics class presentation Ravi S Sinha LIT lab,
In Search of Influential Event Organizers in Online Social Networks
Frequent Subgraph Pattern Mining on Uncertain Graph Data
Integrating Bayesian Networks and Simpson’s Paradox in Data Mining Alex Freitas University of Kent Ken McGarry University of Sunderland.
N EIGHBORHOOD F ORMATION AND A NOMALY D ETECTION IN B IPARTITE G RAPHS Jimeng Sun, Huiming Qu, Deepayan Chakrabarti & Christos Faloutsos Jimeng Sun, Huiming.
Introduction to PageRank Algorithm and Programming Assignment 1 CSC4170 Web Intelligence and Social Computing Tutorial 4 Tutor: Tom Chao Zhou
Estimating the Global PageRank of Web Communities Paper by Jason V. Davis & Inderjit S. Dhillon Dept. of Computer Sciences University of Texas at Austin.
Neighborhood Formation and Anomaly Detection in Bipartite Graphs Jimeng Sun Huiming Qu Deepayan Chakrabarti Christos Faloutsos Speaker: Jimeng Sun.
Subscription Subsumption Evaluation for Content-Based Publish/Subscribe Systems Hojjat Jafarpour, Bijit Hore, Sharad Mehrotra, and Nalini Venkatasubramanian.
Distributed PageRank Computation Based on Iterative Aggregation- Disaggregation Methods Yangbo Zhu, Shaozhi Ye and Xing Li Tsinghua University, Beijing,
CS Lecture 9 Storeing and Querying Large Web Graphs.
© 2011 IBM Corporation IBM Research SIAM-DM 2011, Mesa AZ, USA, Non-Negative Residual Matrix Factorization w/ Application to Graph Anomaly Detection Hanghang.
© 2010 IBM Corporation Diversified Ranking on Large Graphs: An Optimization Viewpoint Hanghang Tong, Jingrui He, Zhen Wen, Ching-Yung Lin, Ravi Konuru.
Zdravko Markov and Daniel T. Larose, Data Mining the Web: Uncovering Patterns in Web Content, Structure, and Usage, Wiley, Slides for Chapter 1:
SCS CMU Proximity Tracking on Time- Evolving Bipartite Graphs Speaker: Hanghang Tong Joint Work with Spiros Papadimitriou, Philip S. Yu, Christos Faloutsos.
Scaling Personalized Web Search Glen Jeh, Jennfier Widom Stanford University Presented by Li-Tal Mashiach Search Engine Technology course (236620) Technion.
Presented by Ozgur D. Sahin. Outline Introduction Neighborhood Functions ANF Algorithm Modifications Experimental Results Data Mining using ANF Conclusions.
Tracking Moving Objects in Anonymized Trajectories Nikolay Vyahhi 1, Spiridon Bakiras 2, Panos Kalnis 3, and Gabriel Ghinita 3 1 St. Petersburg State University.
An introduction to iterative projection methods Eigenvalue problems Luiza Bondar the 23 rd of November th Seminar.
Topic-Sensitive PageRank Taher H. Haveliwala. PageRank Importance is propagated A global ranking vector is pre-computed.
A Local Facility Location Algorithm Supervisor: Assaf Schuster Denis Krivitski Technion – Israel Institute of Technology.
Fast Random Walk with Restart and Its Applications
Abstract Shortest distance query is a fundamental operation in large-scale networks. Many existing methods in the literature take a landmark embedding.
Design Patterns for Efficient Graph Algorithms in MapReduce Jimmy Lin and Michael Schatz University of Maryland MLG, January, 2014 Jaehwan Lee.
GDG DevFest Central Italy Joint work with J. Feldman, S. Lattanzi, V. Mirrokni (Google Research), S. Leonardi (Sapienza U. Rome), H. Lynch (Google)
Ex-MATE: Data-Intensive Computing with Large Reduction Objects and Its Application to Graph Mining Wei Jiang and Gagan Agrawal.
Presented By HaeJoon Lee Yanyan Shen, Beng Chin Ooi, Bogdan Marius Tudor National University of Singapore Wei Lu Renmin University Cang Chen Zhejiang University.
Random Walk with Restart (RWR) for Image Segmentation
Storage and Retrieval of Large RDF Graph Using Hadoop and MapReduce Mohammad Farhan Husain, Pankil Doshi, Latifur Khan, Bhavani Thuraisingham University.
Influence Maximization in Dynamic Social Networks Honglei Zhuang, Yihan Sun, Jie Tang, Jialin Zhang, Xiaoming Sun.
Diversified Top-k Graph Pattern Matching 1 Yinghui Wu UC Santa Barbara Wenfei Fan University of Edinburgh Southwest Jiaotong University Xin Wang.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
ICPP 2012 Indexing and Parallel Query Processing Support for Visualizing Climate Datasets Yu Su*, Gagan Agrawal*, Jonathan Woodring † *The Ohio State University.
1/52 Overlapping Community Search Graph Data Management Lab, School of Computer Science
Xiangnan Kong,Philip S. Yu Multi-Label Feature Selection for Graph Classification Department of Computer Science University of Illinois at Chicago.
Fast Random Walk with Restart and Its Applications Hanghang Tong, Christos Faloutsos and Jia-Yu (Tim) Pan ICDM 2006 Dec , HongKong.
1 Presented by: Yuchen Bian MRWC: Clustering based on Multiple Random Walks Chain.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
A Local Seed Selection Algorithm for Overlapping Community Detection 1 A Local Seed Selection Algorithm for Overlapping Community Detection Farnaz Moradi,
University at BuffaloThe State University of New York Lei Shi Department of Computer Science and Engineering State University of New York at Buffalo Frequent.
1 Panther: Fast Top-K Similarity Search on Large Networks Jing Zhang 1, Jie Tang 1, Cong Ma 1, Hanghang Tong 2, Yu Jing 1, and Juanzi Li 1 1 Department.
Hanghang Tong, Brian Gallagher, Christos Faloutsos, Tina Eliassi-Rad
Powerpoint Templates Page 1 Powerpoint Templates Scalable Text Classification with Sparse Generative Modeling Antti PuurulaWaikato University.
An Efficient Greedy Method for Unsupervised Feature Selection
Graph Data Management Lab, School of Computer Science Branch Code: A Labeling Scheme for Efficient Query Answering on Tree
Melbourne, Australia, Oct., 2015 gSparsify: Graph Motif Based Sparsification for Graph Clustering Peixiang Zhao Department of Computer Science Florida.
By: Jesse Ehlert Dustin Wells Li Zhang Iterative Aggregation/Disaggregation(IAD)
Center-Piece Subgraphs: Problem definition and Fast Solutions Hanghang Tong Christos Faloutsos Carnegie Mellon University.
Progress Report ekker. Problem Definition In cases such as object recognition, we can not include all possible objects for training. So transfer learning.
Optimal Relay Placement for Indoor Sensor Networks Cuiyao Xue †, Yanmin Zhu †, Lei Ni †, Minglu Li †, Bo Li ‡ † Shanghai Jiao Tong University ‡ HK University.
1 / 24 Distributed Methods for High-dimensional and Large-scale Tensor Factorization Kijung Shin (Seoul National University) and U Kang (KAIST)
Dense-Region Based Compact Data Cube
Cohesive Subgraph Computation over Large Graphs
Large Graph Mining: Power Tools and a Practitioner’s guide
Compressive Coded Aperture Video Reconstruction
Privacy Preserving Subgraph Matching on Large Graphs in Cloud
Kijung Shin1 Mohammad Hammoud1
Hanghang Tong, Brian Gallagher, Christos Faloutsos, Tina Eliassi-Rad
Large Graph Mining: Power Tools and a Practitioner’s guide
Probably Approximately
Jinhong Jung, Woojung Jin, Lee Sael, U Kang, ICDM ‘16
Asymmetric Transitivity Preserving Graph Embedding
PRSim: Sublinear Time SimRank Computation on Large Power-Law Graphs.
Presentation transcript:

Kijung Shin Jinhong Jung Lee Sael U Kang BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs Kijung Shin Jinhong Jung Lee Sael U Kang

Introduction Background Experiments Conclusion Proposed Method Introduction How can we measure the relevance (or similarity) between two nodes in a graph? Example: 2 5 1 4 3 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Random Walk with Restart (1) Introduction Background Experiments Conclusion Proposed Method Random Walk with Restart (1) Random Walk with Restart (RWR) assumes a random surfer on a graph 𝑆 𝑆 𝑆 seed node seed node Random walk (with prob 1−𝑐) Restart (with prob 𝑐) BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Random Walk with Restart (2) Introduction Background Experiments Conclusion Proposed Method Random Walk with Restart (2) RWR computes the stationary probability that the surfer stays at each node RWR Score Vector 𝑟 2 5 1 4 3 seed node 0.21 0.31 0.25 0.14 0.09 Node RWR Score (relevance with node 2) 1 0.21 2 - 3 0.14 4 0.25 5 0.09 Restarting probability 𝑐 = 0.2 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Random Walk with Restart (3) Introduction Background Experiments Conclusion Proposed Method Random Walk with Restart (3) Random Walk with Restart (RWR) Goal: measures the relevance between two nodes Properties: accounts for the global network structure and the multi-faceted relationship between nodes Applications: ranking, community detection, link prediction, and anomaly detection BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Research Question Challenge: Question: Introduction Background Experiments Conclusion Proposed Method Research Question Challenge: In many applications, RWR score has to be computed repeatedly with regard to many different seed node 𝑠 Computing RWR from scratch, however, takes too long for large graphs Question: How can we compute RWR on large graphs fast, space efficiently, and accurately? BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Outline Introduction Background <<< Proposed Method Experiments Conclusion BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Introduction Background Experiments Conclusion Proposed Method Problem Definition (1) Given: a graph 𝐺, a seed node 𝑠, and restarting probability 𝑐 Goal: find RWR score vector 𝒓 satisfying 𝒓 = 1−𝑐 𝑨 𝑻 𝒓 +𝑐 𝒒 Probability that the surfer reaches each node by restart Probability that the surfer reaches each node Probability that the surfer reaches each node by random walk where 𝑨 ∈ ℝ 𝒏×𝒏 : row-normalized adjacency matrix 𝒒 ∈ ℝ 𝒏 : query vector where 𝒒 𝒔 =1 and 𝒒 𝒊 =0, ∀𝑖≠𝑠 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Introduction Background Experiments Conclusion Proposed Method Problem Definition (2) Computing RWR boils down to solving a linear system 𝒓 = 1−𝑐 𝑨 𝑻 𝒓 +𝑐 𝒒 ⟺(𝑰− 𝟏−𝒄 𝑨 𝑻 ) 𝒓 =𝑐 𝒒 ⟺ 𝑯 𝒓 =𝑐 𝒒 Given: 𝑯∈ ℝ 𝒏×𝒏 , 𝒒 ∈ ℝ 𝒏 , 𝑐∈ℝ Find: 𝒓 ∈ ℝ 𝒏 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Naïve Method: Inversion (1) Introduction Background Experiments Conclusion Proposed Method Naïve Method: Inversion (1) Goal: With many different 𝒒 s, find 𝒓 satisfying 𝑯 𝒓 =𝑐 𝒒 Preprocess phase (one-time cost): compute 𝑯 −𝟏 Query phase (repetitive cost): Given 𝒒 , compute 𝒓 = 𝑯 −𝟏 (𝑐 𝒒 ) BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Naïve Method: Inversion (2) Introduction Background Experiments Conclusion Proposed Method Naïve Method: Inversion (2) Advantages: Fast query speed (one matrix-vector multiplication) Disadvantages: Inverting 𝑯 takes too long 𝑯 −𝟏 is usually too dense to fit in memory for large graphs Input graph #nz=0.1M (1) Inversion Exact, #nz=527M 𝐻 𝐻 −1 Sparsity pattern of preprocessed matrices on the Routing dataset BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Introduction Background Experiments Conclusion Proposed Method Previous Methods (1) Goal: Replace 𝑯 −𝟏 with sparser matrices by reordering and decomposing 𝑯 Limitation 1: high memory requirements (2) QR decomp. (Fujiwara et al. 12) Exact, #nz=428M (3) LU decomp. (Fujiwara et al. 12) Exact, #nz=10M 𝑄 −1 (= 𝑄 𝑇 ) 𝑅 −1 𝐿 −1 𝑈 −1 Sparsity pattern of preprocessed matrices on the Routing dataset BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Previous Methods (2) Limitation 2: inaccuracy Introduction Background Experiments Conclusion Proposed Method Previous Methods (2) Limitation 2: inaccuracy Challenge: How can we satisfy both space efficiency and accuracy? (4) B_LIN (Tong et al. 07) Approx, #nz=8M (5) NB_LIN (Tong et al. 07) Approx, #nz=3M 𝐴 1 −1 𝑈 𝑉 Λ 𝑈 𝑉 Λ Sparsity pattern of preprocessed matrices on the Routing dataset BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Outline Introduction Background Proposed Method <<< Experiments Conclusion BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

(6) BEAR-Exact (Proposed) Introduction Background Experiments Conclusion Proposed Method Proposed Method: BEAR We propose BEAR, a fast, space-efficient, and accurate RWR computation method (6) BEAR-Exact (Proposed) Exact, #nz=0.4M Method Accurate? #nz Inversion Yes 527M QR 428M LU 10M NB_LIN No 8M B_LIN 3M BEAR-Exact 0.4M 𝐿 2 −1 𝑈 2 −1 𝐿 1 −1 𝑈 1 −1 𝐻 21 𝐻 12 Sparsity pattern of preprocessed matrices on the Routing dataset BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Introduction Background Experiments Conclusion Proposed Method Ideas Behind BEAR (1) In real-world graphs, 𝑯 can be reordered so that it has a large block-diagonal submatrix (Kang et al., 2011) Once 𝑯 is reordered, solving 𝑯 𝒓 =𝑐 𝒒 can be divided into smaller problems (Boyd et al., 2009) 𝑯 𝒓 𝒒 𝐻 Reordered 𝐻 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Ideas Behind BEAR (2) Dividing into smaller problems improves space efficiency and speed Space efficiency can be further improved by applying LU decomposition (Fujiwara et al., 2012) to each of the divided problems Original problem Divided problems Space Requirements BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Introduction Background Experiments Conclusion Proposed Method Outline of BEAR Goal: find 𝒓 satisfying 𝑯 𝒓 =𝑐 𝒒 w.r.t. many different 𝒒 s Preprocessing (one-time cost): Reorder and partition 𝑯 then precomputes several matrices (1) Reordering (2) Partitioning (3) Schur Complement (4) Inverting Query (repetitive cost): Given 𝒒 , compute 𝒓 quickly using the precomputed matrices BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing Phase: (1) Reordering Introduction Background Experiments Conclusion Proposed Method Preprocessing Phase: (1) Reordering Reorder 𝑯 so that it has a large block-diagonal submatrix Utilize hub-and-spoke structure of real-world graphs as in SlashBurn (Kang et al., 2011) hubs spokes real world BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing Phase: (2) Partitioning Introduction Background Experiments Conclusion Proposed Method Preprocessing Phase: (2) Partitioning Partition 𝑯 into four submatrices 𝑯 𝟏𝟏 consists of small diagonal blocks 𝐻 11 real world BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing Phase: (3) Schur Complement Introduction Background Experiments Conclusion Proposed Method Preprocessing Phase: (3) Schur Complement Compute the Schur Complement 𝑺 of 𝑯 𝟏𝟏 𝑺 has the same size of 𝑯 𝟐𝟐 real world BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing Phase: (4) Inverting Introduction Background Experiments Conclusion Proposed Method Preprocessing Phase: (4) Inverting Inverting 𝑺 and each diagonal block of 𝑯 𝟏𝟏 Efficient in terms of time and space because 𝑺 and the diagonal blocks of 𝑯 𝟏𝟏 are small real world BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing Phase: Output Introduction Background Experiments Conclusion Proposed Method Preprocessing Phase: Output Precomputed matrices are small or composed of small diagonal blocks Require little storage real world BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Introduction Background Experiments Conclusion Proposed Method Query Phase Given a query vector 𝒒 , compute the RWR score vector 𝒓 using the precomputed matrices Theorem. Block Elimination (Boyd et al., 2009) : This equation exactly computes RWR scores BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Further Optimization: (1) LU Decomposition Introduction Background Experiments Conclusion Proposed Method Further Optimization: (1) LU Decomposition Instead of inverting 𝑺 and each diagonal block of 𝑯 𝟏𝟏 , invert their LU-decomposed matrices (sparser) Reduce preprocessing time and space requirements 𝐻 11 −1 𝑈 1 −1 𝐿 1 −1 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Further Optimization: (2) Thresholding Introduction Background Experiments Conclusion Proposed Method Further Optimization: (2) Thresholding Drop near-zero entries (if 𝑥 <𝜉) of the precomputed matrices BEAR-Exact: 𝜉=0 No thresholding Guarantee the exactness of RWR BEAR-Approx: 𝜉>0 Reduce query time and space requirements by sacrificing accuracy 𝜉 ↑ ⇒ #non-zeros in precomputed matrices ↓ ⇒ memory efficiency ↑ / query speed ↑ / accuracy ↓ BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Outline Introduction Background Proposed Method Experiments <<< Conclusion BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Experimental Settings Introduction Background Experiments Conclusion Proposed Method Experimental Settings Machine: single PC with with a 4-core CPU and 16GB memory Datasets: large-scale real-world network data BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Competitors Exact methods Approximate methods Inversion Introduction Background Experiments Conclusion Proposed Method Competitors Exact methods Inversion Iterative method LU decomp. (Fujiwara et al., 2012) QR decomp. (Fujiwara et al., 2012) Approximate methods BLIN, NB_LIN (Tong et al., 2008) RPPR, BRPPR (Gleich et al., 2006) BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Space for preprocessed data Introduction Background Experiments Conclusion Proposed Method Q. Space Efficiency How much memory space does BEAR-Exact require for their precomputed matrices? Up to 22x less memory space than competitors Space for preprocessed data BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing time of exact methods Introduction Background Experiments Conclusion Proposed Method Q. Preprocessing Time How long does the preprocessing phase of BEAR- Exact take? Up to 12x less preprocessing time than other methods Preprocessing time of exact methods BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Query time of exact methods Introduction Background Experiments Conclusion Proposed Method Q. Query Time How long does the query phase of BEAR-Exact take? Up to 8x less query time than LU decomp. Up to 300x less query time than Iterative method Query time of exact methods BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Query speed v.s. Accuracy on the Routing dataset Introduction Background Experiments Conclusion Proposed Method Q. Speed vs Accuracy Does BEAR-Approx provide a better trade-off between speed and accuracy than other methods? 250X Query speed v.s. Accuracy on the Routing dataset BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Space for preprocessed data v.s. Accuracy on the Routing dataset Introduction Background Experiments Conclusion Proposed Method Q. Space vs Accuracy Does BEAR-Approx provide a better trade-off between space and accuracy than other methods? 50X Space for preprocessed data v.s. Accuracy on the Routing dataset BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Outline Introduction Background Proposed Method Experiments Conclusion <<< BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Conclusion BEAR (Block Elimination Approach for RWR) BEAR-Exact Introduction Background Experiments Conclusion Proposed Method Conclusion BEAR (Block Elimination Approach for RWR) partitions the adjacency matrix into small submatrices using the hub-and-spoke structure of real-world graphs computes Random Walk with Restart accurately from the submatrices using block elimination BEAR-Exact up to 22× less space, 12× less preprocessing time, and 8× less query time than other exact methods BEAR-Approx better trade-off between time, space, and accuracy than other approximate methods BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Reference Y. Fujiwara, M. Nakatsuji, M. Onizuka, and M. Kitsuregawa. Fast and exact top-k search for random walk with restart. PVLDB, 5(5):442-453, 2012. Y. Fujiwara, M. Nakatsuji, T. Yamamuro, H. Shiokawa, and M. Onizuka. Ecient personalized pagerank with accuracy assurance. In KDD, pages 15-23, 2012. D. Gleich and M. Polito. Approximating personalized pagerank with minimal use of web graph data. Internet Mathematics, 3(3):257-294, 2006 U. Kang and C. Faloutsos. Beyond 'caveman communities': Hubs and spokes for graph compression and mining. In ICDM, pages 300-309, 2011 H. Tong, C. Faloutsos, and J.-Y. Pan. Random walk with restart: fast solutions and applications. KAIS, 14(3):327-346, 2008. S. Boyd and L. Vandenberghe. Convex Optimization. Cambridge University Press, 2009. BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Thank you! The binary codes and datasets are available at http://kdmlab.org/bear BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

SlashBurn (Kang et al. 11) SlashBurn with 𝑘=1 BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Network Structure BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Network Structure BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Time complexity: Space complexity: 𝑂( 𝑖=1 𝑏 𝑛 1𝑖 2 + min 𝑛 1 𝑛 2 ,𝑚 + 𝑛 2 2 ) BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Preprocessing Time BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs

Iterative Method Repeats updating 𝒓 until convergence 𝒓 ← 1−𝑐 𝑨 𝑻 𝒓 +𝑐 𝒒 Advantages: No preprocessing time Easy to implement Disadvantages: Slow (multiple times of matrix-vector multiplication) Inefficient when 𝒓 is computed w.r.t. many different 𝒒 s BEAR: Block Elimination Approach for Random Walk With Restart on Large Graphs