Presentation is loading. Please wait.

Presentation is loading. Please wait.

Analysis of Social Media MLD 10-802, LTI 11-772 William Cohen 10-09-010.

Similar presentations


Presentation on theme: "Analysis of Social Media MLD 10-802, LTI 11-772 William Cohen 10-09-010."— Presentation transcript:

1 Analysis of Social Media MLD 10-802, LTI 11-772 William Cohen 10-09-010

2 Stochastic blockmodel graphs Last week: spectral clustering Theory suggests it will work for graphs produced by a particular generative model Question: can you directly maximize Pr(structure,parameters|data) for that model?

3 Outline Stochastic block models & inference question Review of text models – Mixture of multinomials & EM – LDA and Gibbs (or variational EM) Block models and inference Mixed-membership block models Multinomial block models and inference w/ Gibbs Beastiary of other probabilistic graph models – Latent-space models, exchangeable graphs, p1, ERGM

4 Review – supervised Naïve Bayes Naïve Bayes Model: Compact representation C W1W1 W2W2 W3W3 ….. WNWN C W N  M M   

5 Review – supervised Naïve Bayes Multinomial Naïve Bayes C W1W1 W2W2 W3W3 ….. WNWN  M  For each document d = 1, , M Generate C d ~ Mult( ¢ |  ) For each position n = 1, , N d Generate w n ~ Mult(¢| ,C d )

6 Review – supervised Naïve Bayes Multinomial naïve Bayes: Learning – Maximize the log-likelihood of observed variables w.r.t. the parameters: Convex function: global optimum Solution:

7 Review – unsupervised Naïve Bayes Mixture model: unsupervised naïve Bayes model C W N M   Joint probability of words and classes: But classes are not visible: Z

8 Review – unsupervised Naïve Bayes Mixture model: learning – Not a convex function No global optimum solution – Solution: Expectation Maximization Iterative algorithm Finds local optimum Guaranteed to maximize a lower-bound on the log-likelihood of the observed data

9 Review – unsupervised Naïve Bayes Quick summary of EM: – Log is a concave function – Lower-bound is convex! – Optimize this lower-bound w.r.t. each variable instead X1X1 X2X2 log(0.5x 1 +0.5x 2 ) 0.5log(x 1 )+0.5log(x 2 ) 0.5x 1 +0.5x 2 H(  )

10 Review – unsupervised Naïve Bayes Mixture model: EM solution E-step: M-step: Key capability: estimate distribution of latent variables given observed variables

11 Review - LDA

12 Motivation w M  N Assumptions: 1) documents are i.i.d 2) within a document, words are i.i.d. (bag of words) For each document d = 1, ,M Generate  d ~ D 1 (…) For each word n = 1, , N d generate w n ~ D 2 ( ¢ | θ d n ) Now pick your favorite distributions for D 1, D 2

13 Latent Dirichlet Allocation z w  M  N  For each document d = 1, ,M Generate  d ~ Dir(¢ |  ) For each position n = 1, , N d generate z n ~ Mult( ¢ |  d ) generate w n ~ Mult( ¢ |  z n ) “Mixed membership” K

14 LDA’s view of a document

15 LDA topics

16 Review - LDA Latent Dirichlet Allocation – Parameter learning: Variational EM – Numerical approximation using lower-bounds – Results in biased solutions – Convergence has numerical guarantees Gibbs Sampling – Stochastic simulation – unbiased solutions – Stochastic convergence

17 Review - LDA Gibbs sampling – Applicable when joint distribution is hard to evaluate but conditional distribution is known – Sequence of samples comprises a Markov Chain – Stationary distribution of the chain is the joint distribution Key capability: estimate distribution of one latent variables given the other latent variables and observed variables.

18 Why does Gibbs sampling work? What’s the fixed point? – Stationary distribution of the chain is the joint distribution When will it converge (in the limit)? – Graph defined by the chain is connected How long will it take to converge? – Depends on second eigenvector of that graph

19

20 Called “collapsed Gibbs sampling” since you’ve marginalized away some variables Fr: Parameter estimation for text analysis - Gregor Heinrich

21 Review - LDA Latent Dirichlet Allocation z w  M  N  Randomly initialize each z m,n Repeat for t=1,…. For each doc m, word n Find Pr(z mn =k|other z’s) Sample z mn according to that distr. “Mixed membership”

22 Outline Stochastic block models & inference question Review of text models – Mixture of multinomials & EM – LDA and Gibbs (or variational EM) Block models and inference Mixed-membership block models Multinomial block models and inference w/ Gibbs Beastiary of other probabilistic graph models – Latent-space models, exchangeable graphs, p1, ERGM

23 Statistical Models of Networks Want a generative probabilistic model that’s amenable to analysis…. … but more expressive than Erdos-Renyi One approach: exchangeable graph model – Exchangeable: X1,X2 are exchangable if Pr(X1,X2,W)=Pr(X2,X1,W). – The generalizes of i.i.d.-ness – It’s a Bayesian thing

24 Review - LDA Motivation w M  N Assumptions: 1) documents are i.i.d 2) within a document, words are i.i.d. (bag of words) For each document d = 1, ,M Generate  d ~ D 1 (…) For each word n = 1, , N d generate w n ~ D 2 ( ¢ | θ d n ) Docs and words are exchangeable.

25 Stochastic Block models: assume 1) nodes w/in a block z and 2) edges between blocks z p,z q are exchangeable zpzp zqzq a pq N2N2 zpzp N  p 

26 Stochastic Block models: assume 1) nodes w/in a block z and 2) edges between blocks z p,z q are exchangeable zpzp zqzq a pq N2N2 zpzp N  p  Gibbs sampling: Randomly initialize z p for each node p. For t = 1… For each node p Compute z p given other z’s Sample z p See: Snijders & Nowicki, 1997, Estimation and Prediction for Stochastic Blockmodels for Groups with Latent Graph Structure

27 Mixed Membership Stochastic Block models pp qq zp.zp. z.qz.q a pq N2N2 pp N  p  Airoldi et al, JMLR 2008

28 Mixed Membership Stochastic Block models

29

30 Parkkinen et al paper

31 Another mixed membership block model

32 z=(zi,zj) is a pair of block ids n z = #pairs z q z1, i = #links to i from block z1 q z1,. = #outlinks in block z1 δ = indicator for diagonal M = #nodes

33 Another mixed membership block model

34

35 Outline Stochastic block models & inference question Review of text models – Mixture of multinomials & EM – LDA and Gibbs (or variational EM) Block models and inference Mixed-membership block models Multinomial block models and inference w/ Gibbs Beastiary of other probabilistic graph models – Latent-space models, exchangeable graphs, p1, ERGM

36 Exchangeable Graph Model Defined by a 2 k x 2 k table q(b 1,b 2 ) Draw a length-k bit string b(n) like 01101 for each node n from a uniform distribution. For each pair of node n,m – Flip a coin with bias q(b(n),b(m)) – If it’s heads connect n,m complicated Pick k-dimensional vector u from a multivariate normal w/ variance α and covariance β – so u i ’s are correlated. Pass each u i thru a sigmoid so it’s in [0,1] – call that p i Pick b i using p i

37 Exchangeable Graph Model Pick k-dimensional vector u from a multivariate normal w/ variance α and covariance β – so u i ’s are correlated. Pass each u i thru a sigmoid so it’s in [0,1] – call that p i Pick b i using p i If α is big then ux,uy are really big (or small) so px,py will end up in a corner. 01 1

38 The p 1 model for a directed graph Parameters, per node i: – Θ: background edge probability – α i : “expansiveness” – how extroverted is i? – β i : “ popularity ” – how much do others want to be with i? – ρ i : “reciprocation” – how likely is i to respond to an incomping link with an outgoing one? Logistic-regression like procedure can be used to fit this to data from a graph

39 Exponential Random Graph Model Basic idea: – Define some features of the graph (e.g., number of edges, number of triangles, …) – Build a MaxEnt-style model based on these features

40 Latent Space Model Each node i has a latent position in Euclidean space, z(i) z(i)’s drawn from a mixture of Gaussians Probability of interaction between i and j depend on the distance between z(i) and z(j) Inference is a little more complicated… [Handcock & Raftery, 2007]

41

42

43 Outline Stochastic block models & inference question Review of text models – Mixture of multinomials & EM – LDA and Gibbs (or variational EM) Block models and inference Mixed-membership block models Multinomial block models and inference w/ Gibbs Beastiary of other probabilistic graph models – Latent-space models, exchangeable graphs, p1, ERGM


Download ppt "Analysis of Social Media MLD 10-802, LTI 11-772 William Cohen 10-09-010."

Similar presentations


Ads by Google