790 likes | 808 Views
Explore the power of Bregman divergences in clustering and dimensionality reduction, moving beyond Euclidean distance to improve algorithms. Learn about k-means, PCA, and more with a focus on probabilistic interpretations.
E N D
Bregman Divergences in Clustering and Dimensionality ReductionCOMS 6998-4: Learning and Empirical Inference Irina Rish IBM T.J. Watson Research Center Slide credits: Srujana Merugu, Arindam Banerjee, Sameer Agarwal
Outline • Intro to Bregman Divergences • Clustering with Bregman Divergences • k-means: quick overview • From Euclidean distance to Bregman divergences • Some rate-distortion theory • Dimensionality Reduction with Bregman Divergences • PCA: quick overview • Probabilistic Interpretation of PCA; exponential family • From Euclidean distance to Bregman divergences • Conclusions
Distance (distortion) measures in learning • Euclidean distance – most commonly used • Nearest neighbor, k-means clustering, least squares regression, PCA, distance metric learning, etc • But…is it always an appropriate type of distance? No! • Nominal attributes (e.g. binary) • Distances between distributions • Probabilistic interpretation: • Euclidean distance Gaussian data • Beyond Gaussian? Exponential family distributions Bregman divergences
Squared Euclidean distance is a Bregman divergence
Relative entropy (i.e., KL-divergence) is another Bregman divergence
Now, how about generalizing soft clustering Algorithms using Bregman divergences?