Panos Ipeirotis Stern School of Business New York University Analyzing User-Generated Content using Econometrics.

Slides:



Advertisements
Similar presentations
Mobile and Internet Systems Group Reputation Premiums in Electronic Peer-to-Peer Markets: Analyzing Textual Feedback and Network Structure Presenter: Sean.
Advertisements

1 Modeling Query-Based Access to Text Databases Eugene Agichtein Panagiotis Ipeirotis Luis Gravano Computer Science Department Columbia University.
Panos Ipeirotis Stern School of Business New York University Analyzing User-Generated Content using Econometrics.
Answering Approximate Queries over Autonomous Web Databases Xiangfu Meng, Z. M. Ma, and Li Yan College of Information Science and Engineering, Northeastern.
Temporal Query Log Profiling to Improve Web Search Ranking Alexander Kotov (UIUC) Pranam Kolari, Yi Chang (Yahoo!) Lei Duan (Microsoft)
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Imbalanced data David Kauchak CS 451 – Fall 2013.
1.Accuracy of Agree/Disagree relation classification. 2.Accuracy of user opinion prediction. 1.Task extraction performance on Bing web search log with.
Big Data Stupid Decisions The Importance Of Measuring What We Should Be Measuring Stern School of Business, New York University “A Computer.
Discriminant Analysis Database Marketing Instructor:Nanda Kumar.
By: Mr Hashem Alaidaros MIS 211 Lecture 4 Title: Data Base Management System.
Show me the Money! Deriving the Pricing Power of Product Features by Mining Consumer Reviews. Nikolay Archak, Anindya Ghose, Panagiotis Ipeirotis New York.
Anindya Ghose Panos Ipeirotis Arun Sundararajan Stern School of Business New York University Opinion Mining using Econometrics A Case Study on Reputation.
Prof. Panos Ipeirotis Search and the New Economy Session 5 Mining User-Generated Content.
Evaluating Search Engine
Search Engines and Information Retrieval
Software Quality Metrics
Data Mining By Archana Ketkar.
CONTENT-BASED BOOK RECOMMENDING USING LEARNING FOR TEXT CATEGORIZATION TRIVIKRAM BHAT UNIVERSITY OF TEXAS AT ARLINGTON DATA MINING CSE6362 BASED ON PAPER.
Multimedia Data Mining Arvind Balasubramanian Multimedia Lab The University of Texas at Dallas.
Virtual Business: Retailing
Get Another Label? Using Multiple, Noisy Labelers Joint work with Victor Sheng and Foster Provost Panos Ipeirotis Stern School of Business New York University.
Get Another Label? Improving Data Quality and Data Mining Using Multiple, Noisy Labelers Panos Ipeirotis Stern School of Business New York University Joint.
Panos Ipeirotis Stern School of Business New York University Structuring and querying online opinions using econometrics.
Designing Ranking Systems for Hotels on Travel Search Engines by Mining User-Generated and Crowd sourced Content Author - Anindya Ghose, Panagiotis G.
Nikolay Archak,Anindya Ghose,Panagiotis G. Ipeirotis Class Presentation By: Arunava Bhattacharya.
Walter Hop Web-shop Order Prediction Using Machine Learning Master’s Thesis Computational Economics.
Data Mining : Introduction Chapter 1. 2 Index 1. What is Data Mining? 2. Data Mining Functionalities 1. Characterization and Discrimination 2. MIning.
Mining and Summarizing Customer Reviews
Data Mining Techniques
Panos Ipeirotis Stern School of Business New York University Opinion Mining using Econometrics A Case Study on Reputation Systems Join work with Anindya.
Opinion Mining Using Econometrics: A Case Study on Reputation Systems Anindya Ghose, Panagiotis G. Ipeirotis, and Arun Sundararajan Leonard N. Stern School.
© 2013 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Panos Ipeirotis Stern School of Business New York University Opinion Mining Using Econometrics.
Panos Ipeirotis New York University Opinion Mining using Econometrics A Case Study on Reputation Systems Joint work with Anindya Ghose and Arun Sundararajan.
Chapter 6 : Software Metrics
1 A Bayesian Method for Guessing the Extreme Values in a Data Set Mingxi Wu, Chris Jermaine University of Florida September 2007.
1 Bins and Text Categorization Carl Sable (Columbia University) Kenneth W. Church (AT&T)
Pete Bohman Adam Kunk. What is real-time search? What do you think as a class?
Designing Ranking Systems for Consumer Reviews: The Economic Impact of Customer Sentiment in Electronic Markets Anindya Ghose Panagiotis Ipeirotis Stern.
Chapter 8 - slide 1 Copyright © 2009 Pearson Education, Inc. Publishing as Prentice Hall Chapter Eight Product, Services, and Brands Building Customer.
Marketing CHAPTER Marketing Basics
Panos Ipeirotis New York University Opinion Mining using Econometrics A Case Study on Reputation Systems Joint work with Anindya Ghose and Arun Sundararajan.
Direct Teacher: Professor Ng Reporter: Cindy Pineapple 1 Summarized from :
Panos Ipeirotis Stern School of Business New York University Text Mining of Electronic News Content for Economic Research “On the Record”: A Forum on Electronic.
Panos Ipeirotis Stern School of Business New York University Opinion Mining using Econometrics A Case Study on Reputation Systems Join work with Anindya.
1 Rated Aspect Summarization of Short Comments Yue Lu, ChengXiang Zhai, and Neel Sundaresan.
1 Rated Aspect Summarization of Short Comments Yue Lu, ChengXiang Zhai, and Neel Sundaresan Presented by: Sapan Shah.
1. 2 Traditional Income Statement LO1: Prepare a contribution margin income statement.
Chapter 07 Customer-Driven Marketing Strategy: Creating Value for Target Customers.
3-1 Data Mining Kelby Lee. 3-2 Overview ¨ Transaction Database ¨ What is Data Mining ¨ Data Mining Primitives ¨ Data Mining Objectives ¨ Predictive Modeling.
Market Research & Product Management.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
A Metrics Program. Advantages of Collecting Software Quality Metrics Objective assessments as to whether quality requirements are being met can be made.
McCormick Northwestern Engineering 1 Electrical Engineering & Computer Science Mining Millions of Reviews: A Technique to Rank Products Based on Importance.
Information Extraction Lecture 3 – Rule-based Named Entity Recognition Dr. Alexander Fraser, U. Munich September 3rd, 2014 ISSALE: University of Colombo.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
Show Me the Money! Deriving the Pricing Power of Product Features by Mining Consumer Reviews Nikolay Archak, Anindya Ghose, and Panagiotis G. Ipeirotis.
Lesson 16: Analyzing the Competition. Objectives Determine who the competition is and how they are affecting your business Determine who the competition.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Chapter 3 Building Business Intelligence Chapter 3 DATABASES AND DATA WAREHOUSES Building Business Intelligence 6/22/2016 1Management Information Systems.
Data Mining – Introduction (contd…) Compiled By: Umair Yaqub Lecturer Govt. Murray College Sialkot.
Marketing Foundations What is Marketing? What is the goal of Marketing?
Opinion spam and Analysis 소프트웨어공학 연구실 G 최효린 1 / 35.
UNIT – V BUSINESS ANALYTICS
Erasmus University Rotterdam
Text Analysis and Search Analytics
Conjoint analysis.
Text Analysis and Search Analytics
Presentation transcript:

Panos Ipeirotis Stern School of Business New York University Analyzing User-Generated Content using Econometrics

Comparative Shopping

Are Customers Irrational? $11.04 (+1.5%) BuyDig.com gets Price Premium (customers pay more than the minimum price)

Price Premiums / Amazon Are Buyers Irrational (?) (paying more) Are Sellers Irrational (?) (charging less)

Why not Buying the Cheapest? You buy more than a product  Customers do not pay only for the product  Customers also pay for a set of fulfillment characteristics  Delivery  Packaging  Responsiveness  … Customers care about reputation of sellers!

Example of a reputation profile

The Idea in a Single Slide Conjecture: Price premiums measure reputation Reputation is captured in text feedback Our contribution: Examine how text affects price premiums (and learn to rank opinion phrases as a side effect) ACL 2007

Data: Variables of Interest Price Premium  Difference of price charged by a seller minus listed price of a competitor Price Premium = (Seller Price – Competitor Price)  Calculated for each seller-competitor pair, for each transaction  Each transaction generates M observations, (M: number of competing sellers) Alternative Definitions:  Average Price Premium (one per transaction)  Relative Price Premium (relative to seller price)  Average Relative Price Premium (combination of the above)

Decomposing Reputation Is reputation just a scalar metric?  Previous studies assumed a “monolithic” reputation  Decompose reputation in individual components  Sellers characterized by a set of fulfillment characteristics (packaging, delivery, and so on) What are these characteristics (valued by consumers?)  We think of each characteristic as a dimension, represented by a noun, noun phrase, verb or verbal phrase (“shipping”, “packaging”, “delivery”, “arrived”)  We scan the textual feedback to discover these dimensions

Decomposing and Scoring Reputation Decomposing and scoring reputation  We think of each characteristic as a dimension, represented by a noun or verb phrase (“shipping”, “packaging”, “delivery”, “arrived”)  The sellers are rated on these dimensions by buyers using modifiers (adjectives or adverbs), not numerical scores  “Fast shipping!”  “Great packaging”  “Awesome unresponsiveness”  “Unbelievable delays”  “Unbelievable price” How can we find out the meaning of these adjectives?

Structuring Feedback Text: Example What is the reputation score of this feedback? P1: I was impressed by the speedy delivery! Great Service! P2: The item arrived in awful packaging, but the delivery was speedy Deriving reputation score  We assume that a modifier assigns a “score” to a dimension  α(μ, k): score associated when modifier μ evaluates the k-th dimension  w(k): weight of the k-th dimension  Thus, the overall (text) reputation score Π(i) is a sum: Π(i) =2*α (speedy, delivery)* weight(delivery)+ 1*α (great, service)* weight(service) + 1*α (awful, packaging)* weight(packaging) unknown unknown?

Sentiment Scoring with Regressions Scoring the dimensions  Use price premiums as “true” reputation score Π(i)  Use regression to assess scores (coefficients) Regressions  Control for all variables that affect price premiums  Control for all numeric scores of reputation  Examine effect of text: E.g., seller with “fast delivery” has premium $10 over seller with “slow delivery”, everything else being equal  “fast delivery” is $10 better than “slow delivery” estimated coefficients Π(i) =2*α (speedy, delivery)* weight(delivery)+ 1*α (great, service)* weight(service) + 1*α (awful, packaging)* weight(packaging) Price Premium

Measuring Reputation Regress textual reputation against price premiums Example for “delivery”: –Fast delivery vs. Slow delivery: +$7.95 –So “fast” is better than “slow” by a $7.95 margin

Some Indicative Dollar Values Positive Negative Natural method for extracting sentiment strength and polarity good packaging -$0.56 Naturally captures the pragmatic meaning within the given context captures misspellings as well Positive? Negative ?

More Results Further evidence: Who will make the sale?  Classifier that predicts sale given set of sellers  Binary decision between seller and competitor  Used Decision Trees (for interpretability)  Training on data from Oct-Jan, Test on data from Feb-Mar  Only prices and product characteristics: 55%  + numerical reputation (stars), lifetime: 74%  + encoded textual information: 89%  text only: 87% Text carries more information than the numeric metrics

Looking Back Comprehensive setting –All information about merchants stored at feedback profile Easy text processing –Large number of feedback postings (100’s and 1000’s of postings common) –Short and concise language

Similar Setting: Word of “Mouse” Consumer reviews –Derived from user experience –Describe different product features –Provide subjective evaluations of product features Product reviews affect product sales –What is the importance of each product feature? –What is the consumer evaluation of each feature? Apply the same techniques? I love virtually everything about this camera....except the lousy picture quality. The camera looks great, feels nice, is easy to use, starts up quickly, and is of course waterproof. It fits easily in a pocket and the battery lasts for a reasonably long period of time.

Contrast with Reputation Significant data sparseness Smaller number of reviews per product –Typically reviews vs ,000 postings Much longer than feedback postings –2-3 paragraphs each, vs characters in reputation Not an isolated system Consumers form opinions from many sources

Bayesian Learning Approach Consumers perform Bayesian learning of product attributes using signals from reviews –Consumers have prior expectations of quality –Consumers update expectation from new signals

Online shopping as learning Belief for Image Quality Updated Belief for Image Quality “excellent image quality” “fantastic image quality” “superb image quality” “great image quality” “fantastic image quality” “superb image quality” Updated Belief for Image Quality  Consumers pick the product that maximizes their expected utility

Expected Utility Consumers pick the product that maximizes their expected utility Expected utility based on: –Mean of the evaluation and –Uncertainty of the evaluation Notice: negative reviews may increase sales! DesignImage Quality + U= Mean(design) Var(design) Mean(img qual) Variance(img qual)

Examine changes in demand and infer parameters Product Reviews and Product Sales “poor lens” +3% “excellent lens” -1% “poor photos” +6% “excellent photos” - 2%  Feature “photos” is two time more important than “lens”  “Excellent” is positive, “poor” is negative  “Excellent” is three times stronger than “poor”

Feature Weights for Digital Cameras SLRPoint & Shoot

New Product Search Approach Consumers want the “best product” first Best product: Highest value for the money –Maximize (gained) product utility –Minimize (lost) utility of money

Utility of Money The highest the available income, the lowest the utility of money (i.e., rich people spend easier)

Hotel Search Application Transaction data from big travel search website Computed “expected utility” for each hotel using: –Reviews –Satellite photos for landscape (beach, downtown, highway,…) –Location statistics (crime, etc) and points of interest Substracted “utility of money” based on its price Ranked according to “consumer surplus” (i.e., difference of two)

Hotel Ranking Percentage of users preferring econometric ranking in blind comparison Cities Tripadvisor Travelocity Price Low to high Price High to Low Hotel Class # Review # of Amenit ies New York72%68%62%70%66% 62% Los Angeles68%66%64%84%88%80%64% San Francisco 84%62%68%72%66% 70% Orlando 64%68% 74%66%74%66% New Orleans 84%62%83%64% 66%70% Salt Lake City80% 68%72%82%66%68% Significance Level P=0.01 ≥ 66% P=0.001 ≥ 72% (Sign Test, N=50)

Other Applications Financial news and price/variance prediction Measuring (and predicting) importance of political events Deriving better keyword bidding, pricing, and ad generation strategies

Other Projects SQoUT project Structured Querying over Unstructured Text Managing Noisy Labelers Amazon Mechanical Turk, “Wisdom of the Crowds” 31

32 SQoUT: Structured Querying over Unstructured Text Information extraction applications extract structured relations from unstructured text May , Atlanta -- The Centers for Disease Control and Prevention, which is in the front line of the world's response to the deadly Ebola epidemic in Zaire, is finding itself hard pressed to cope with the crisis… DateDisease NameLocation Jan. 1995MalariaEthiopia July 1995Mad Cow DiseaseU.K. Feb. 1995PneumoniaU.S. May 1995EbolaZaire Information Extraction System (e.g., NYU’s Proteus) Disease Outbreaks in The New York Times

33 SQoUT: The Questions Output Tokens … Extraction System(s) Text Databases 3.Extract output tuples 2.Process documents 1.Retrieve documents from database/web/archive Questions: 1.How to we retrieve the documents? 2.How to configure the extraction systems? 3.What is the execution time? 4.What is the output quality? SIGMOD’06, TODS’07, ICDE’09, TODS’09

Mechanical Turk Example 34

Motivation Labels can be used in training predictive models  Duplicate detection systems  Image recognition  Web search But: labels obtained from above sources are noisy. This directly affects the quality of learning models  How can we know the quality of annotators?  How can we know the correct answer?  How can we use best noisy annotators?

36 Quality and Classification Performance Labeling quality increases  classification quality increases Q = 0.5 Q = 0.6 Q = 0.8 Q = 1.0

Tradeoffs for Classification Get more labels  Improve label quality  Improve classification Get more examples  Improve classification Q = 0.5 Q = 0.6 Q = 0.8 Q = 1.0 KDD 2008

Thank you! Questions?

Price Amazon

Average price Amazon

Relative Price Premiums

Average Relative Price Premiums

Capturing transactions and “price premiums” Data: Transactions Seller ListingItemPrice When item is sold, listing disappears

Capturing transactions and “price premiums” Data: Transactions While listing appears, item is still available time Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10

Capturing transactions and “price premiums” Data: Transactions While listing appears, item is still available time Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10 Item still not sold on 1/7

Capturing transactions and “price premiums” Data: Transactions When item is sold, listing disappears time Item sold on 1/9 Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10

Seller: uCameraSite.com 1.Canon Powershot x300 2.Kodak - EasyShare 5.0MP 3.Nikon - Coolpix 5.1MP 4.Fuji FinePix Canon PowerShot x900 Reputation Pricing Tool for Sellers Your last 5 transactions in Cameras Name of productPrice Seller 1 - $431 Seller 2 - $409 You - $399 Seller 3 - $382 Seller 4-$379 Seller 5-$376 Canon Powershot x300 Your competitive landscape Product Price ( reputation ) (4.8) (4.65) (4.7) (3.9) (3.6) (3.4) Your Price: $399 Your Reputation Price: $419 Your Reputation Premium: $20 (5%) $20 Left on the table

Quantitatively Understand & Manage Seller Reputation RSI Tool for Seller Reputation Management How your customers see you relative to other sellers: 35%* 69% 89% 82% 95% Service Packaging Delivery Overall Quality Dimensions of your reputation and the relative importance to your customers: Service Packaging Delivery Quality Other * Percentile of all merchants RSI Products Automatically Identify the Dimensions of Reputation from Textual Feedback Dimensions are Quantified Relative to Other Sellers and Relative to Buyer Importance Sellers can Understand their Key Dimensions of Reputation and Manage them over Time Arms Sellers with Vital Info to Compete on Reputation Dimensions other than Low Price.

Marketplace Search Buyer’s Tool Used Market (ex: Amazon) Price Range $250-$300 Seller 1Seller 2 Seller 4Seller 3 Sort by Price/Service/Delivery/other dimensions Canon PS SD700 Service Packaging Delivery Price Dimension Comparison Seller 1 PriceServicePackageDelivery Seller 2 Seller 3 Seller 4 Seller 5 Seller 6 Seller 7

Data Overview  Panel of 280 software products sold by Amazon.com X 180 days  Data from “used goods” market  Amazon Web services facilitate capturing transactions  We do not use any proprietary Amazon data (Details in the paper)

Data: Secondary Marketplace

Data: Capturing Transactions time Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 We repeatedly “crawl” the marketplace using Amazon Web Services While listing appears  item is still available  no sale

Data: Capturing Transactions time Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10 We repeatedly “crawl” the marketplace using Amazon Web Services When listing disappears  item sold

Weights of Hotel Characteristics Based on Different Travel Purposes Consumers with different travel purposes assign different weight distributions on the same set of hotel characteristics.

Sensitivity to Rating and Review Count Based on Different Age Groups Age pay more attention to online reviews compared to other age groups.

User Study Experiment 1: Blind pair-wise comparisons, 100 anonymous AMT users; 8 existing baselines: -Price low to high -Price high to low -Online review count -Hotel class -Hotel size (number of rooms -Number of internal amenities -TripAdvisor reviewer rating -Travelocity reviewer rating Conclusion: CS-based ranking is overwhelmingly preferred. Reasoning: Diversity, satisfies consumers’ multidimensional preferences

User Study Experiment 2: Blind pair-wise comparisons, 100 anonymous AMT users; baseline: generalized CS-based ranking (for an average consumer). E.g., Business trip and family trip AMT user study results in the NYC experiment. Conclusion: Personalized CS-based ranking is overwhelmingly preferred. Reasoning: Capture consumers’ specific expectations, dovetail with their real purchase motivation.

Estimation Results Capture Real Motivation e.g., Business travelers indicated that they prefer quiet inner environment and easy access to highway or public transportation. This was fully captured in our estimation results, see (b).