1 / 22

Analysis of Social Media MLD 10-802, LTI 11-772

Analysis of Social Media MLD 10-802, LTI 11-772. William Cohen 10- 16- 010. Review - LDA. “Mixed membership”. Latent Dirichlet Allocation. . Randomly initialize each z m,n Repeat for t=1,…. For each doc m, word n Find Pr( z mn = k |other z’s)

lauren
Download Presentation

Analysis of Social Media MLD 10-802, LTI 11-772

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Analysis of Social MediaMLD 10-802, LTI 11-772 William Cohen 10-16-010

  2. Review - LDA “Mixed membership” • Latent Dirichlet Allocation  • Randomly initialize each zm,n • Repeat for t=1,…. • For each doc m, word n • Find Pr(zmn=k|other z’s) • Sample zmn according to that distr. a z w N M 

  3. Outline • Stochastic block models & inference question • Review of text models • Mixture of multinomials & EM • LDA and Gibbs (or variational EM) • Block models and inference • Mixed-membership block models • Multinomial block models and inference w/ Gibbs • Beastiary of other probabilistic graph models • Latent-space models, exchangeable graphs, p1, ERGM

  4. Parkkinen et al paper

  5. Another mixed membership block model

  6. Another mixed membership block model z=(zi,zj) is a pair of block ids nz = #pairs z qz1,i = #links to i from block z1 qz1,. = #outlinks in block z1 δ = indicator for diagonal M = #nodes

  7. Another mixed membership block model

  8. Another mixed membership block model

  9. Outline • Stochastic block models & inference question • Review of text models • Mixture of multinomials & EM • LDA and Gibbs (or variational EM) • Block models and inference • Mixed-membership block models • Multinomial block models and inference w/ Gibbs • Beastiary of other probabilistic graph models • Latent-space models, exchangeable graphs, p1, ERGM

  10. Latent Space Model • Each node i has a latent position in Euclidean space, z(i) • z(i)’s drawn from a mixture of Gaussians • Probability of interaction between i and j depend on the distance between z(i) and z(j) • Inference is a little more complicated… [Handcock & Raftery, 2007]

  11. Airoldi’s MMSBM

  12. Outline • Stochastic block models & inference question • Review of text models • Mixture of multinomials & EM • LDA and Gibbs (or variational EM) • Block models and inference • Mixed-membership block models • Multinomial block models and inference w/ Gibbs • Beastiary of other probabilistic graph models • Latent-space models, exchangeable graphs, p1, ERGM

  13. Exchangeable Graph Model • Defined by a 2k x 2k table q(b1,b2) • Draw a length-k bit string b(n) like 01101 for each node n from a uniform distribution. • For each pair of node n,m • Flip a coin with bias q(b(n),b(m)) • If it’s heads connect n,m complicated • Pick k-dimensional vector u from a multivariate normal w/ variance α and covariance β – so ui’s are correlated. • Pass each uithru a sigmoid so it’s in [0,1] – call that pi • Pick biusing pi

  14. Exchangeable Graph Model 1 If α is big then ux,uy are really big (or small) so px,py will end up in a corner. • Pick k-dimensional vector u from a multivariate normal w/ variance α and covariance β – so ui’s are correlated. • Pass each ui thru a sigmoid so it’s in [0,1] – call that pi • Pick biusing pi 0 1

  15. Exchangeable Graph Model 1 If α is big then ux,uy are really big (or small) so px,py will end up in a corner. • Pick k-dimensional vector u from a multivariate normal w/ variance α and covariance β – so ui’s are correlated. • Pass each uithru a sigmoid so it’s in [0,1] – call that pi • Pick biusing pi 0 1

  16. The p1 model for a directed graph • Parameters, per node i: • Θ: background edge probability • αi: “expansiveness” – how extroverted is i? • βi:“popularity”– how much do others want to be with i? • ρij: “reciprocation” – how likely is itorespond to an incomping link with an outgoing one? + ρij Logistic-regression like procedure can be used to fit this to data from a graph

  17. Exponential Random Graph Model • Basic idea: • Define some features of the graph (e.g., number of edges, number of triangles, …) • Build a MaxEnt-style model based on these features • General: • includes Erdos-Renyi, p1, … • Issues • Partition function is intractible • Alternative: model conditional pseudo-likelihood of a each edge (i.e., Pr(edge|rest of graph)

  18. Kroneker product graphs

  19. Kroneker product graphs

  20. Kroneker product graphs • Good fit to many commonly-observed network properties • scale-free degree distribution • diameter • … • Gradient descent can be used to fit an “initiator matrix” to a real adjacency matrix

More Related