180 likes | 294 Views
Kernel Stick-Breaking Process. D. B. Dunson and J. Park. Discussion led by Qi An Jan 19 th , 2007. Outline. Motivation Model formulation and properties Prediction rules Posterior Computation Examples Conclusions. Motivation.
E N D
Kernel Stick-Breaking Process D. B. Dunson and J. Park Discussion led by Qi An Jan 19th, 2007
Outline • Motivation • Model formulation and properties • Prediction rules • Posterior Computation • Examples • Conclusions
Motivation • Consider a problem of estimating the conditional density of a response variable using a mixture model, , where Gx is an unknown probability measure indexed by x. • The problem of defining priors for random probability measures on Gx has received increasing attention in recent year. For example, DP, DDP.
One model • In DDP, the atoms can vary with x according to a stochastic process while the weights are fixed • Dunson et al propose a model to allow the weights to vary with predictors while this model lacks reasonable marginalization and updating properties.
Model formulation • Introduce a countable sequence of mutually independent random components • The kernel stick-breaking process (KSBP) can be defined as follows:
About the model • The model for Gx is a predictor-dependent mixture over an infinite sequence of basis probability measures, Gh* located at Γh. • Bases located close to x and having a smaller index, h, tend to receive higher probability weight. • KSBP accommodates dependency between Gx and Gx’
Special cases • If K(x,Γ)=1 for all and Gh*~DP(αG0), it is a stick-breaking mixture of DP. • If K(x,Γ)=1, and , we obtain Gx≡G, with G having a stick-breaking prior. • If and , we obtain a Pitman-Yor process.
Properties • Let , we can obtain • The correlation between measures First moment No dependency on V and Γ Second moment It can be proven and the value 1 in the limit as x x’ where
Alternative representation The KSBP has an alternative representation The moments and correlation coefficient has the form
Truncation • For stick-breaking Gibbs sampler, we need to make truncation approximation • Author proves that the residual weights decrease exponentially fast in N and an accurate approximation may be obtained for moderate N The approximated model can be expressed as
Prediction rules • Consider a special case in which The model can be equivalently expressed as:
Prediction rules • Define and is a subset of the integers between 1 and n • It can be proven that the probability that subjects i and j belong to the same cluster is The predictive distribution is obtained by marginalization where and denote the set of possible r- dimensional subsets of {1,…,s} that include i
Posterior Computation From the prior, we can obtain 1, sample Si 2, sample CSi when Si=0 (assign subject I to a new atom at an occupied location) 3, sample θh
4, sample Vh 5, sample Γh using a Metropolis-Hastings step or Gibbs step if H is a set of discrete potential locations First sample and then, alternate between (i) Sampling (Aih,Bih) from their conditional distribution (ii) Updating Vh by sampling from conditional posterior
Conclusions • This stick-breaking process is useful in setting in which there is uncertainty in an uncountable collection of probability measures • The process can be applied in predictor dependent clustering, dynamic modeling and spatial data analysis, besides the density regression. • The KSBP formulation can be applied to many tools developed for exchangeable stick-breaking processes with minimal modification. • A predicator dependent urn scheme is obtained, which generalizes the Polya urn scheme