Loading…

Coresets for kernel clustering

We devise coresets for kernel k - Means with a general kernel, and use them to obtain new, more efficient, algorithms. Kernel k - Means has superior clustering capability compared to classical k - Means , particularly when clusters are non-linearly separable, but it also introduces significant compu...

Full description

Saved in:
Bibliographic Details
Published in:Machine learning 2024-08, Vol.113 (8), p.5891-5906
Main Authors: Jiang, Shaofeng H. -C., Krauthgamer, Robert, Lou, Jianing, Zhang, Yubo
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We devise coresets for kernel k - Means with a general kernel, and use them to obtain new, more efficient, algorithms. Kernel k - Means has superior clustering capability compared to classical k - Means , particularly when clusters are non-linearly separable, but it also introduces significant computational challenges. We address this computational issue by constructing a coreset, which is a reduced dataset that accurately preserves the clustering costs. Our main result is a coreset for kernel k - Means that works for a general kernel and has size poly ( k ϵ - 1 ) . Our new coreset both generalizes and greatly improves all previous results; moreover, it can be constructed in time near-linear in n . This result immediately implies new algorithms for kernel k - Means , such as a ( 1 + ϵ ) -approximation in time near-linear in n , and a streaming algorithm using space and update time poly ( k ϵ - 1 log n ) . We validate our coreset on various datasets with different kernels. Our coreset performs consistently well, achieving small errors while using very few points. We show that our coresets can speed up kernel K - M E A N S + + (the kernelized version of the widely used K - M E A N S + + algorithm), and we further use this faster kernel K - M E A N S + + for spectral clustering. In both applications, we achieve significant speedup and a better asymptotic growth while the error is comparable to baselines that do not use coresets.
ISSN:0885-6125
1573-0565
DOI:10.1007/s10994-024-06540-z