Exploiting Social Context for Review Quality Prediction Yue Lu University of Illinois at Urbana-Champaign Panayiotis Tsaparas Microsoft Research Alexandros.

Slides:



Advertisements
Similar presentations
A Comparison of Implicit and Explicit Links for Web Page Classification Dou Shen 1 Jian-Tao Sun 2 Qiang Yang 1 Zheng Chen 2 1 Department of Computer Science.
Advertisements

Regularization David Kauchak CS 451 – Fall 2013.
On-line learning and Boosting
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Random Projection for High Dimensional Data Clustering: A Cluster Ensemble Approach Xiaoli Zhang Fern, Carla E. Brodley ICML’2003 Presented by Dehong Liu.
Logistic Regression.
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Partitioned Logistic Regression for Spam Filtering Ming-wei Chang University of Illinois at Urbana-Champaign Wen-tau Yih and Christopher Meek Microsoft.
Pattern Recognition and Machine Learning
Learning to Rank: New Techniques and Applications Martin Szummer Microsoft Research Cambridge, UK.
1 Transfer Learning Algorithms for Image Classification Ariadna Quattoni MIT, CSAIL Advisors: Michael Collins Trevor Darrell.
Presented by Li-Tal Mashiach Learning to Rank: A Machine Learning Approach to Static Ranking Algorithms for Large Data Sets Student Symposium.
Predicting the Semantic Orientation of Adjective Vasileios Hatzivassiloglou and Kathleen R. McKeown Presented By Yash Satsangi.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
MSRC Summer School - 30/06/2009 Cambridge – UK Hybrids of generative and discriminative methods for machine learning.
Heterogeneous Consensus Learning via Decision Propagation and Negotiation Jing Gao † Wei Fan ‡ Yizhou Sun † Jiawei Han † †University of Illinois at Urbana-Champaign.
Heterogeneous Consensus Learning via Decision Propagation and Negotiation Jing Gao† Wei Fan‡ Yizhou Sun†Jiawei Han† †University of Illinois at Urbana-Champaign.
OCFS: Optimal Orthogonal Centroid Feature Selection for Text Categorization Jun Yan, Ning Liu, Benyu Zhang, Shuicheng Yan, Zheng Chen, and Weiguo Fan et.
Semi-Supervised Clustering Jieping Ye Department of Computer Science and Engineering Arizona State University
Collaborative Ordinal Regression Shipeng Yu Joint work with Kai Yu, Volker Tresp and Hans-Peter Kriegel University of Munich, Germany Siemens Corporate.
Graph-Based Semi-Supervised Learning with a Generative Model Speaker: Jingrui He Advisor: Jaime Carbonell Machine Learning Department
Distributed Representations of Sentences and Documents
Scalable Text Mining with Sparse Generative Models
Maria-Florina Balcan A Theoretical Model for Learning from Labeled and Unlabeled Data Maria-Florina Balcan & Avrim Blum Carnegie Mellon University, Computer.
Adapting Deep RankNet for Personalized Search
Online Learning Algorithms
STRUCTURED PERCEPTRON Alice Lai and Shi Zhi. Presentation Outline Introduction to Structured Perceptron ILP-CRF Model Averaged Perceptron Latent Variable.
Collaborative Filtering Matrix Factorization Approach
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
More Machine Learning Linear Regression Squared Error L1 and L2 Regularization Gradient Descent.
Transfer Learning From Multiple Source Domains via Consensus Regularization Ping Luo, Fuzhen Zhuang, Hui Xiong, Yuhong Xiong, Qing He.
Opinion Integration and Summarization Yue Lu University of Illinois at Urbana-Champaign.
Learning with Positive and Unlabeled Examples using Weighted Logistic Regression Wee Sun Lee National University of Singapore Bing Liu University of Illinois,
Group Recommendations with Rank Aggregation and Collaborative Filtering Linas Baltrunas, Tadas Makcinskas, Francesco Ricci Free University of Bozen-Bolzano.
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
Semisupervised Learning A brief introduction. Semisupervised Learning Introduction Types of semisupervised learning Paper for review References.
8/25/05 Cognitive Computations Software Tutorial Page 1 SNoW: Sparse Network of Winnows Presented by Nick Rizzolo.
Xiaoxiao Shi, Qi Liu, Wei Fan, Philip S. Yu, and Ruixin Zhu
Confidence-Aware Graph Regularization with Heterogeneous Pairwise Features Yuan FangUniversity of Illinois at Urbana-Champaign Bo-June (Paul) HsuMicrosoft.
Mining Social Network for Personalized Prioritization Language Techonology Institute School of Computer Science Carnegie Mellon University Shinjae.
Mining Social Networks for Personalized Prioritization Shinjae Yoo, Yiming Yang, Frank Lin, II-Chul Moon [KDD ’09] 1 Advisor: Dr. Koh Jia-Ling Reporter:
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
Page 1 Inferring Relevant Social Networks from Interpersonal Communication Munmun De Choudhury, Winter Mason, Jake Hofman and Duncan Watts WWW ’10 Summarized.
Exploit of Online Social Networks with Community-Based Graph Semi-Supervised Learning Mingzhen Mo and Irwin King Department of Computer Science and Engineering.
WSP: A Network Coordinate based Web Service Positioning Framework for Response Time Prediction Jieming Zhu, Yu Kang, Zibin Zheng and Michael R. Lyu The.
CoCQA : Co-Training Over Questions and Answers with an Application to Predicting Question Subjectivity Orientation Baoli Li, Yandong Liu, and Eugene Agichtein.
Multiple Instance Learning for Sparse Positive Bags Razvan C. Bunescu Machine Learning Group Department of Computer Sciences University of Texas at Austin.
ICONIP 2010, Sydney, Australia 1 An Enhanced Semi-supervised Recommendation Model Based on Green’s Function Dingyan Wang and Irwin King Dept. of Computer.
Optimal Reverse Prediction: Linli Xu, Martha White and Dale Schuurmans ICML 2009, Best Overall Paper Honorable Mention A Unified Perspective on Supervised,
Discriminative Training and Machine Learning Approaches Machine Learning Lab, Dept. of CSIE, NCKU Chih-Pin Liao.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Meta-Path-Based Ranking with Pseudo Relevance Feedback on Heterogeneous Graph for Citation Recommendation By: Xiaozhong Liu, Yingying Yu, Chun Guo, Yizhou.
Ariel Fuxman, Panayiotis Tsaparas, Kannan Achan, Rakesh Agrawal (2008) - Akanksha Saxena 1.
Applying Link-based Classification to Label Blogs Smriti Bhagat, Irina Rozenbaum Graham Cormode.
Big Data Infrastructure Week 8: Data Mining (1/4) This work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States.
Semi-Supervised Clustering
Deep Feedforward Networks
Linli Xu Martha White Dale Schuurmans University of Alberta
Boosting and Additive Trees
Collaborative Filtering Matrix Factorization Approach
CSCI B609: “Foundations of Data Science”
Learning Emoji Embeddings Using Emoji Co-Occurrence Network Graph
RECOMMENDER SYSTEMS WITH SOCIAL REGULARIZATION
Graph-based Security and Privacy Analytics via Collective Classification with Joint Weight Learning and Propagation Binghui Wang, Jinyuan Jia, and Neil.
Recursively Adapted Radial Basis Function Networks and its Relationship to Resource Allocating Networks and Online Kernel Learning Weifeng Liu, Puskal.
Jonathan Elsas LTI Student Research Symposium Sept. 14, 2007
Logistic Regression.
Semi-Supervised Learning
Label propagation algorithm
Presentation transcript:

Exploiting Social Context for Review Quality Prediction Yue Lu University of Illinois at Urbana-Champaign Panayiotis Tsaparas Microsoft Research Alexandros Ntoulas Microsoft Research Livia Polanyi Microsoft April 28, WWW’2010 Raleigh, NC

2 Why do we care about Predicting Review Quality? User reviews (1764) User “helpfulness” votes help prioritize reading User “helpfulness” votes help prioritize reading But not all reviews have votes 1.New reviews 2.Reviews aggregated from multiple sources But not all reviews have votes 1.New reviews 2.Reviews aggregated from multiple sources

What has been done? 3 As classification or regression problem √ × ? ? ? ? ? ? ? ? ? √ [Zhang&Varadarajan`06] [Kim et al. `06] [Liu et al. `08] [Ghose&Ipeirotis `10] Labeled Unlabeled Textual features Meta-data features

Reviews are NOT Stand-Alone Documents We also observe… 4 ReviewerIdentitySocialNetwork Social Context =+ Our Work: Exploiting Social Context for Review Quality Prediction Our Work: Exploiting Social Context for Review Quality Prediction

Roadmap Motivation Review Quality Prediction Algorithms Experimental Evaluation Conclusions 5

SentiPositive SentiNegative Text-only Baseline 6 Textual Features Text Statistics NumSent NumTokens SentLen CapRatio UniqWordRatio Syntactic POS:RB POS:PP POS:V POS:CD POS:JJ POS:NN POS:SYM POS:COM POS:FW Conformity KLDiv Sentiment FeatureVector( )=

Base Model: Linear Regression 7 w = argmin = argmin{ } Quality( ) = Weights×FeatureVector( ) i i Closed-form: w=

Straight-forward Approach: Adding Social Context as Features 8 Reviewer History NumReview AvgRating Social Network InDegree OutDegree PageRank Textual Features Social Context Features FeatureVector( )= Disadvantages: Social context features not always available Anonymous reviews? A new reviewer? Need more training data Disadvantages: Social context features not always available Anonymous reviews? A new reviewer? Need more training data

Our Approach: Social Context as Constraints 9 ReviewerIdentity SocialNetwork Quality( ) is related to Quality( ) is related to its Social Network Our Intuitions: How to combine such intuitions with Textual info?

Formally: Graph-based Regularizers 10 { + β× Graph Regularizer } w = argmin Trade-off parameter Designed to “favor” our intuitions Baseline Loss function Advantages: Semi-supervised: make use of unlabeled data Applicable to reviews without social context LabeledUnlabeled We will define four regularizers base on four hypotheses.

1.Reviewer Consistency Hypothesis 11 Quality( ) Quality( ) ~ Quality( ) 2 Quality( ) ~ 3 Reviewers are consistent! Reviewers are consistent!

Regularizer for Reviewer Consistency 12 Reviewer Regularizer =∑ [ Quality( ) - Quality( ) ] 2 Quality( ) ] Sum over all data (train + test) for all pairs reviews in the same-author graph Closed-form solution! Same-Author Graph (A) [Zhou et al. 03] [Zhu et al. 03] [Belkin et al 06] w= Graph LaplacianReview-Feature Matrix

2.Trust Consistency Hypothesis 13 Quality( ) - Quality( ) ≤ 0 I trust people with quality at least as good as mine! AVG ( Quality( ) ) Defined as

Regularizer for Trust Consistency 14 Trust Regularizer =∑max[0, Quality( ) - Quality( )] 2 Quality( )] 2 Sum over all data (train + test) for all pairs of reviewers connected in the trust graph No closed-form solution… Still convex  Gradient Descent Trust Graph

3.Co-Citation Consistency Hypothesis 15 Quality( ) - Quality( ) → 0 Trust GraphCo-citation Graph I am consistent with my “trust standard”!

Regularizer for Co-citation Consistency 16 Co-citation Regularizer =∑[ Quality( ) - Quality( ) ] 2 Quality( ) ] 2 Closed-form solution! Sum over all data (train + test) for all pairs of reviewers connected in the co-citation graph Co-citation Graph (C) w= Review-Reviewer Matrix

4.Link Consistency Hypothesis 17 Quality( ) - Quality( ) → 0 Trust GraphLink Graph I trust people with similar quality as mine!

Regularizer for Link Consistency 18 Link Regularizer =∑[ Quality( ) - Quality( ) ] 2 Quality( ) ] 2 Closed-form solution! Sum over all data (train + test) for all pairs of reviewers connected in the co-citation graph Link Graph

Roadmap Motivation Review Quality Prediction Algorithms Experimental Evaluation Conclusions 19

Data from Ciao UK StatisticsCellphoneBeautyDigital Camera # Reviews Reviews/Reviewer ratio Trust Graph Density SummaryCellphoneBeautyDigital Camera Social Contextrich sparse Gold-std Quality Distribution balancedskewedbalanced

Hypotheses Testing: Reviewer Consistency 21 Qg( ) - 1 Qg( ) 2 Qg( ) - 1 Qg( ) 3 Reviewer Consistency Hypothesis supported by data Difference in Review Quality Density From same reviewer From different reviewers (Cellphone)

Hypotheses Testing: Social Network-based Consistencies 22 Qg( ) - Qg( ) B is not linked to A B trusts A B is co-cited with A B is linked to A BA Social Network-based Consistencies supported by data Difference in Reviewer Quality Density (Cellphone)

Prediction Performance: Exploiting Social Context 23 % of MSE Difference Percentage of Training Data 10%25%50%100% AddFeatures is most effective given sufficient training data With limited training data, Reg methods work best With limited training data, Reg methods work best Reg:Reviewer > Reg:Trust > Reg:Cocitation > Reg:Link (Cellphone) Better Reg:Link AddFeatures Reg:Reviewer Reg:Cocitation Reg:Trust

Prediction Performance: Compare Three Categories 24 % of MSE Difference CellphoneBeauty Digital Camera Better Reg:Link Reg:Reviewer Reg:Cocitation Reg:Trust Improvement on Digital Camera is smaller due to sparse social context Reviews/Reviewer ratio = 1.06

Parameter Sensitivity 25 Text-only Baseline (Cellphone)(Beauty) Regularization Parameter Mean Squared Error consistently better than Baseline when parameter < 0.1 Better

Conclusions Improve Review Quality Prediction using Social Context Formalize into a Semi-supervised Graph Regularization framework Utilize both labeled and unlabeled data Applicable on data with no social context Promising results on real world data – Esp. limited labels, rich social context 26

Future Work Combine multiple regularizers Optimize by nDCG instead of MSE Infer trust network Spam detection 27

Thank you! & Questions?