Download presentation

Presentation is loading. Please wait.

Published byBruno McCormick Modified over 3 years ago

1
Consumer Behavior Prediction using Parametric and Nonparametric Methods Elena Eneva CALD Masters Presentation 19 August 2002 Advisors: Alan Montgomery, Rich Caruana, Christos Faloutsos

2
Outline Introduction Data Economics Overview Baseline Models New Hybrid Models Results Conclusions and Future Work

3
Background Retail chains are aiming to customize prices in individual stores Pricing strategies should adapt to the neighborhood demand Stores can increase operating profit margins by 33% to 83%

4
Price Elasticity consumer’s response to price change inelasticelastic Q is quantity purchased P is price of product

5
Data Example

6
Data Example – Log Space

7
Assumptions Independence –Substitutes: fresh fruit, other juices –Other Stores Stationarity –Change over time –Holidays

8
“The” Model Category Price of Product 1 Price of Product 2 Price of Product 3 Price of Product N... “I know your customers” Predictor Quantity bought of Product 1... Quantity bought of Product 2 Quantity bought of Product 3 Quantity bought of Product N Need to multiply this across many stores, many categories. convert to ln spaceconvert to original space

9
Converting to Original Space

10
Existing Methods Traditionally – using parametric models (linear regression) Recently – using non-parametric models (neural networks)

11
Our Goal Advantage of LR: known functional form (linear in log space), extrapolation ability Advantage of NN: flexibility, accuracy robustness accuracy NN new LR Take Advantage: use the known functional form to bias the NN Build hybrid models from the baseline models

12
Datasets weekly store-level cash register data at the product level Chilled Orange Juice category 2 years 12 products 10 random stores selected

13
Evaluation Measure Root Mean Squared Error (RMS) the average deviation between the predicted quantity and the true quantity

14
Models Hybrids –Smart Prior –MultiTask Learning –Jumping Connections –Frozen Jumping Connections Baselines –Linear Regression –Neural Networks

15
Baselines Linear Regression Neural Networks

16
q is the quantity demanded p i is the price for the i th product K products overall The coefficients a and b i are determined by the condition that the sum of the square residuals is as small as possible. Linear Regression

18
Results RMS

19
Neural Networks generic nonlinear function approximators a collection of basic units (neurons), computing a (non)linear function of their input backpropagation

20
Neural Networks 1 hidden layer, 100 units, sigmoid activation function

21
Results RMS

22
Hybrids Smart Prior MultiTask Learning Jumping Connections Frozen Jumping Connections

23
Smart Prior Idea: start the NN at a “good” set of weights, help it start from a “smart” prior. Take this prior from the known “linearity” NN first trained on synthetic data generated by the LR model NN then trained on the real data

24
Smart Prior

25
Results RMS

26
Multitask Learning Idea: learning an additional related task in parallel, using a shared representation Adding the output of the LR model (built over the same inputs) as an extra output to the NN Make the net share its hidden nodes between both tasks Custom halting function Custom RMS function

27
MultiTask Learning

28
Results RMS

29
Jumping Connections Idea: fusing LR and NN change architecture add connections which “jump” over the hidden layer Gives the effect of simulating a LR and NN all together

30
Jumping Connections

31
Results RMS

32
Frozen Jumping Connections Idea: you have the linearity, now use it! same architecture as Jumping Connections, plus really emphasizing the linearity freeze the weights of the jumping layer, so the network can’t “forget” about the linearity

33
Frozen Jumping Connections

36
Results RMS

37
Models Hybrids –Smart Prior –MultiTask Learning –Jumping Connections –Frozen Jumping Connections Baselines: –Linear Regression –Neural Networks Combinations –Voting –Weighted Average

38
Combining Models Idea: Ensemble Learning Committee Voting – equal weights for each model’s prediction Weighted Average – optimal weights determined by a linear regression model 2 baseline and 3 hybrid models (Smart Prior, MultiTask Learning, Frozen Jumping Conections)

39
Committee Voting Average the predictions of the models

40
Results RMS

41
Weighted Average – Model Regression Linear regression on baselines and hybrid models to determine vote weights

42
Results RMS

43
Normalized RMS Error Compare model performance across stores Stores of different sizes, ages, locations, etc Need to normalize Compare to baselines Take the error of the LR benchmark as unit error

44
Normalized RMS Error

45
Conclusions Clearly improved models for customer choice prediction Will allow stores to price the products more strategically and optimize profits Maintain better inventories Understand product interaction

46
Future Work Ideas analyze Weighted Average model compare extrapolation ability of new models use other domain knowledge –shrinkage model – a “super” store model with data pooled across all stores

47
Acknowledgements I would like to thank my advisors and my CALDling friends and colleagues

48
The Most Important Slide for this presentation and the paper: www.cs.cmu.edu/~eneva/research.htm eneva@cs.cmu.edu

49
References Montgomery, A. (1997). Creating Micro- Marketing Pricing Strategies Using Supermarket Scanner Data West, P., Brockett, P. and Golden, L (1997) A Comparative Analysis of Neural Networks and Statistical Methods for Predicting Consumer Choice Guadagni, P. and Little, J. (1983) A Logit Model of Brand Choice Calibrated on Scanner data Rossi, P. and Allenby, G. (1993) A Bayesian Approach to Estimating Household Parameters

50
Error Measure – Unbiased Model Details which is an unbiased estimator for q. is a biased estimator for q, so we correct the bias by using by computing the integral over the distribution

51
On one hand… In log space, Price-Quantity relationship is fairly linear

52
On the other hand… the derivation of consumers' demand responses to price changes without the need to write down and rely upon particular mathematical models for demand

Similar presentations

Presentation is loading. Please wait....

OK

Sparse vs. Ensemble Approaches to Supervised Learning

Sparse vs. Ensemble Approaches to Supervised Learning

© 2018 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on javascript events textbox Ppt on fire extinguisher types list Ppt on kotak life insurance Ppt on statistics and probability review Ppt on european union Ppt on national parks and sanctuaries in india Ppt on earth dam design Ppt on amplitude shift keying spectrum Ppt on health insurance in india Ppt on coalition government canada