Edge Weight Prediction in Weighted Signed Networks

Slides:



Advertisements
Similar presentations
Node labels as random variables prior belief observed neighbor potentials compatibility potentials Opinion Fraud Detection in Online Reviews using Network.
Advertisements

Absorbing Random walks Coverage
Learning to Recommend Hao Ma Supervisors: Prof. Irwin King and Prof. Michael R. Lyu Dept. of Computer Science & Engineering The Chinese University of Hong.
Propagation of Trust and Distrust Antti Sorjamaa Propagation of Trust and Distrust R. Guha, R. Kumar, P. Raghavan and A. Tomkins New York, 2004 Antti Sorjamaa.
The loss function, the normal equation,
Lecture 9 Measures and Metrics. Structural Metrics Degree distribution Average path length Centrality Degree, Eigenvector, Katz, Pagerank, Closeness,
Computing Trust in Social Networks
October 28, 2010Neural Networks Lecture 13: Adaptive Networks 1 Adaptive Networks As you know, there is no equation that would tell you the ideal number.
Structure, Tie Persistence and Event Detection in Large Phone and SMS Networks Leman Akoglu and Bhavana Dalvi {lakoglu, Carnegie Mellon.
DATA MINING LECTURE 13 Absorbing Random walks Coverage.
Predicting Positive and Negative Links in Online Social Networks
Mining Social Network for Personalized Prioritization Language Techonology Institute School of Computer Science Carnegie Mellon University Shinjae.
Mining Social Networks for Personalized Prioritization Shinjae Yoo, Yiming Yang, Frank Lin, II-Chul Moon [KDD ’09] 1 Advisor: Dr. Koh Jia-Ling Reporter:
Today Ensemble Methods. Recap of the course. Classifier Fusion
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
EigenRank: A Ranking-Oriented Approach to Collaborative Filtering IDS Lab. Seminar Spring 2009 강 민 석강 민 석 May 21 st, 2009 Nathan.
1 A fast algorithm for learning large scale preference relations Vikas C. Raykar and Ramani Duraiswami University of Maryland College Park Balaji Krishnapuram.
PREDIcT: Towards Predicting the Runtime of Iterative Analytics Adrian Popescu 1, Andrey Balmin 2, Vuk Ercegovac 3, Anastasia Ailamaki
Zibin Zheng DR 2 : Dynamic Request Routing for Tolerating Latency Variability in Cloud Applications CLOUD 2013 Jieming Zhu, Zibin.
Detecting Communities Via Simultaneous Clustering of Graphs and Folksonomies Akshay Java Anupam Joshi Tim Finin University of Maryland, Baltimore County.
Poorva Potdar Sentiment and Textual analysis of Create-Debate data EECS 595 – End Term Project.
Computer Science 1 Using Clustering Information for Sensor Network Localization Haowen Chan, Mark Luk, and Adrian Perrig Carnegie Mellon University
Single-Pass Belief Propagation
Jure Leskovec (Stanford), Daniel Huttenlocher and Jon Kleinberg (Cornell)
Supervised Random Walks: Predicting and Recommending Links in Social Networks Lars Backstrom (Facebook) & Jure Leskovec (Stanford) Proc. of WSDM 2011 Present.
Importance Measures on Nodes Lecture 2 Srinivasan Parthasarathy 1.
Optimizing Average Precision using Weakly Supervised Data Aseem Behl 1, C.V. Jawahar 1 and M. Pawan Kumar 2 1 IIIT Hyderabad, India, 2 Ecole Centrale Paris.
Reputation-aware QoS Value Prediction of Web Services Weiwei Qiu, Zhejiang University Zibin Zheng, The Chinese University of HongKong Xinyu Wang, Zhejiang.
Data Science Credibility: Evaluating What’s Been Learned
Correlation Clustering
Negative Link Prediction and Its Applications in Online Political Networks Mehmet Yigit Yildirim Mert Ozer Hasan Davulcu.
Finding Dense and Connected Subgraphs in Dual Networks
Uncovering the Mystery of Trust in An Online Social Network
The PageRank Citation Ranking: Bringing Order to the Web
SLAQ: Quality-Driven Scheduling for Distributed Machine Learning
New Characterizations in Turnstile Streams with Applications
Semi-supervised Machine Learning Gergana Lazarova
Comparison of Social Networks by Likhitha Ravi
Empirical analysis of Chinese airport network as a complex weighted network Methodology Section Presented by Di Li.
Predicting Long-Term Impact of CQA Posts: A Comprehensive Viewpoint
Predicting Positive and Negative Links in Online Social Networks
CS 4/527: Artificial Intelligence
Logistic Regression Classification Machine Learning.
Intro to Machine Learning
More Graph Algorithms.
Networks with Signed Edges
COS 518: Advanced Computer Systems Lecture 12 Mike Freedman
Introduction to Statistics
Evaluation and Its Methods
Predict Failures with Developer Networks and Social Network Analysis
CS539: Project 3 Zach Pardos.
Jinhong Jung, Woojung Jin, Lee Sael, U Kang, ICDM ‘16
RECOMMENDER SYSTEMS WITH SOCIAL REGULARIZATION
Ensembles.
Binghui Wang, Le Zhang, Neil Zhenqiang Gong
Asymmetric Transitivity Preserving Graph Embedding
Intro to Machine Learning
The loss function, the normal equation,
Graph-based Security and Privacy Analytics via Collective Classification with Joint Weight Learning and Propagation Binghui Wang, Jinyuan Jia, and Neil.
Neural Networks II Chen Gao Virginia Tech ECE-5424G / CS-5824
GANG: Detecting Fraudulent Users in OSNs
Mathematical Foundations of BME Reza Shadmehr
Junheng, Shengming, Yunsheng 11/09/2018
Evaluation and Its Methods
Dynatrace AI Demystified
Neural Networks II Chen Gao Virginia Tech ECE-5424G / CS-5824
David Kauchak CS158 – Spring 2019
Evaluation and Its Methods
Machine Learning: Lecture 5
Logistic Regression Geoff Hulten.
Presentation transcript:

Edge Weight Prediction in Weighted Signed Networks Srijan Kumar, Univ. of Maryland Francesca Spezzano, Boise State Univ. V.S. Subrahmanian, Univ. of Maryland Christos Faloutsos, Carnegie Mellon Univ.

Ratings are everywhere Ratings are everywhere. On platforms like Amazon and Yelp, people rate products from 1-5 stars. On other platforms like Epinions, Slashdot and Bitcoin trust networks, people rate other people to express opinion towards each other.

Weighted Signed Edges Positive edges: Trust, Like, Support, Agree -0.2 +0.1 -0.9 +1 Positive edges: Trust, Like, Support, Agree Negative edges: Distrust, Dislike, Oppose, Disagree Weights: Strength of the relation On any such platform, relations can be expressed very naturally using weighted and signed edges. For instance, a person may like or trust another a lot or a little, or dislike someone a lot and someone else a little. All these relations - trust, like, support, agreement - can be represented with positive edges and distrust, dislike, opposition and disagreement can be represented with negative edges. The strength of these relations can be represented by their weight. Without loss of generality, we say that the edge weights lie between -1 and +1. These relations can be between user-user or user-product. Edge weights lie between -1 and +1

Predicting Edge Weight ? ? ? How to accurately predict weight and sign of missing edges? So the task in this work is to predict edge weights of missing edges. So given a network with some weighted and signed edges, can we predict the weights and signs of the edges that are not visible? To answer these questions, we develop two metrics called fairness and goodness. Our Solution: Fairness and Goodness

Example -0.5 0.8 1.0 1.0 -1.0 -0.9 -0.95 Add weights Entity being rated does not need to be a person, can be a product Why is goodness in [-1, 1], why fairness is [0,1] -1.0 -0.9 -0.95

Intuition: Fairness and Goodness Fairness: how reliable a user is in rating others Fairness f(u) ∈ [0,1] A user is fair if it gives “correct” ratings to other users. Goodness: how fair user rate it Goodness g(v) ∈ [-1, 1] A user is good if it gets high ratings from fair users. Very general, not specific to WSNs

Goodness W(u,v) Fairness f(u) Goodness g(v) Weighted incoming rating

Fairness W(u,v) Fairness Goodness f(u) g(v) Average deviation of user u’s ratings Deviation of rating from goodness

Fairness and Goodness Algorithm Initialization Update Goodness Update Fairness

Initialization: All Fair and All Good f(u) = 1 f(u) = 1 g(v) = 1 f(u) = 1 g(v) = 1 f(u) = 1 g(v) = 1 f(u) = 1 f(u) = 1

Updating Goodness - Iteration 1 f(u) = 1 f(u) = 1 g(v) = 0.67 g(v) = -0.67 f(u) = 1 f(u) = 1 f(u) = 1 f(u) = 1

Updating Fairness - Iteration 1 f(u) = 0.58 f(u) = 0.92 g(v) = 0.67 f(u) = 0.92 g(v) = 0.67 f(u) = 0.92 g(v) = -0.67 f(u) = 0.92 f(u) = 0.92

… repeat until convergence f(u) = 0.17 f(u) = 0.83 g(v) = 0.67 f(u) = 0.83 g(v) = 0.67 f(u) = 0.83 g(v) = -0.67 f(u) = 0.83 f(u) = 0.83 Predicted Edge Weight (u,v) = f(u) x g(v)

Theoretical Guarantees Convergence Theorem: The error between iterations is bounded, and as t increases, the rating scores converge. The error bound is given by: As t increases, Uniqueness Theorem: Iterations converge to a unique solution, given the starting criteria. Time Complexity: O(|E|)

Experiments: Data and settings Two Bitcoin trust networks: trust/distrust. 6k nodes, 36k edges Wikipedia editor: agree/disagree 342k nodes, 5.6M edges User-user network: like/dislike 365k nodes, 2.6M edges Bitcoin Trust datasets: Trust/Distrust Wikipedia Request for Adminship: Trust/Distrust, Like/Dislike Wikipedia Editor Network: Agree/Disagree Epinions: Trust/Distrust, Agree/Disagree Twitter: Like/Dislike, Agree/Disagree Wikipedia adminship: support/oppose 10k nodes, 100k edges User-user network: trust/distrust 196k nodes, 4.8M edges

Comparisons Reciprocal edge weight Triadic balance theory Triadic status theory Local status theory Weighted PageRank Signed Eigenvector Centrality Signed-HITS Bias and Deserve TidalTrust Algorithm EigenTrust Algorithm MDS Algorithm Methods are adapted to work on weighted and signed networks, whenever applicable. Performance metrics: Root Mean Square Error (RMSE) and Pearson Correlation Coefficient (PCC)

Prediction: Leave-one-edge-out Deleted weight of edge Two predictions: Predicted edge weight (u,v) = g(v) Predicted edge weight (u,v) = f(u) x g(v) Fairness and Goodness predictions are overall more accurate than existing algorithms

Prediction: Supervised Regression Model Prediction by all methods are put into a regression model and trained on the training edge set. The learned model is used to predict edge weight of test edge. Fairness and Goodness features are the most important features in the Linear Regression model in most networks. ADD feature importance

Prediction: N% Edge Removal Lower error is better Fairness and Goodness performs the best

Prediction: N% Edge Removal Higher correlation is better Fairness and Goodness performs the best

Conclusions Two novel metrics: Fairness and Goodness General metrics for any weighted graph In this work, used to predict edge weight in weighted signed networks Scalable, with time complexity O(|E|) Guaranteed solution Performs the best in predicting edge weights, both under leave-one-edge-out and N% edge removal cross-validation

Thank you! Datasets and code at: http://cs.umd.edu/~srijan/wsn Reach me at: srijan@cs.umd.edu Website: http://cs.umd.edu/~srijan

Applications Identify potential customers Add new aspect to standard graph mining tasks: Node ranking Anomaly detection Clustering Community detection Sentiment prediction Information diffusion