100 likes | 231 Views
The Phylogenetic Indian Buffet Process : A Non-Exchangeable Nonparametric Prior for Latent Features By: Kurt T. Miller, Thomas L. Griffiths and Michael I. Jordan ICML 2008. Presented by: John Paisley Duke University, ECE. Motivation.
E N D
The Phylogenetic Indian Buffet Process: A Non-Exchangeable Nonparametric Prior for Latent FeaturesBy: Kurt T. Miller, Thomas L. Griffiths and Michael I. JordanICML 2008 Presented by: John Paisley Duke University, ECE
Motivation • Nonparametric models are often used with the assumption of exchangeability. • The Indian Buffet Process is an example • Sometimes, non-exchangeable models might be more appropriate. • The Phylogenetic Indian Buffet Process • Similar to the IBF, but uses additional information of how related diners are with each other. • These relationships are captured in a tree structure.
Phylogenetic Indian Buffet Process • Uses a tree to model columns zk • This is done as follows: • Assign the root node to be zero • Along an edge of distance t, let this change to a 1 with probability , where . The distance from every leaf to the root is 1. • If a 0 is changed to a 1 along a path to a node, all subsequent nodes are 1 and therefore so are the leaves.
Sampling Issues • For (1), use the sum-product algorithm (Pearl, 1988). • For (2), use the chain rule of probability. • An MCMC inference algorithm is given in detail.
Experimental Results • Elimination by Aspects (EBA) model • A Choice Model • Let there be i objects and zik indicate the ith object has the kth feature. Let each feature have a weight, wk. The EBA model defines the probability of choosing object I over j as • The likelihood of an observation matrix, X, is • This has been modeled using the IBP.
Experimental Results • Consider now an underlying tree structure to this model. • Preference trees: Out of 9 personalities, 3 movie stars, 3 athletes and 3 politicians, people made the 36 pairwise choices of whom they would rather spend time with. Here, L is the length of the edge of each general category to a leaf. • A soft version of this tree is modeled with the pIBP using data generated from this model with L = 0.1
Experimental Results • Example results: As the number of samples decreases, the pIBP is able to infer the structure better than the IBP because of the prior.
Experimental Results • As can be seen, the additional structure in the model produces better results.