1 / 79

Irina Rish IBM T.J. Watson Research Center

Bregman Divergences in Clustering and Dimensionality Reduction COMS 6998-4: Learning and Empirical Inference. Irina Rish IBM T.J. Watson Research Center. Slide credits: Srujana Merugu, Arindam Banerjee, Sameer Agarwal. Outline. Intro to Bregman Divergences

hshanna
Download Presentation

Irina Rish IBM T.J. Watson Research Center

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Bregman Divergences in Clustering and Dimensionality ReductionCOMS 6998-4: Learning and Empirical Inference Irina Rish IBM T.J. Watson Research Center Slide credits: Srujana Merugu, Arindam Banerjee, Sameer Agarwal

  2. Outline • Intro to Bregman Divergences • Clustering with Bregman Divergences • k-means: quick overview • From Euclidean distance to Bregman divergences • Some rate-distortion theory • Dimensionality Reduction with Bregman Divergences • PCA: quick overview • Probabilistic Interpretation of PCA; exponential family • From Euclidean distance to Bregman divergences • Conclusions

  3. Distance (distortion) measures in learning • Euclidean distance – most commonly used • Nearest neighbor, k-means clustering, least squares regression, PCA, distance metric learning, etc • But…is it always an appropriate type of distance? No! • Nominal attributes (e.g. binary) • Distances between distributions • Probabilistic interpretation: • Euclidean distance  Gaussian data • Beyond Gaussian? Exponential family distributions Bregman divergences

  4. Squared Euclidean distance is a Bregman divergence

  5. Relative entropy (i.e., KL-divergence) is another Bregman divergence

  6. Recall Bregman Diverences

  7. Now, how about generalizing soft clustering Algorithms using Bregman divergences?

  8. (natural parameter)

More Related