English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

A Sober Look at Clustering Stability

MPS-Authors
There are no MPG-Authors in the publication available
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

Ben-David, S., von Luxburg, U., & Pal, D. (2006). A Sober Look at Clustering Stability. In G. Lugosi, & H. Simon (Eds.), Learning Theory: 19th Annual Conference on Learning Theory, COLT 2006, Pittsburgh, PA, USA, June 22-25, 2006 (pp. 5-19). Berlin, Germany: Springer.


Cite as: https://hdl.handle.net/11858/00-001M-0000-0013-D03D-2
Abstract
Stability is a common tool to verify the validity of sample based algorithms. In clustering it is widely used to tune the parameters of
the algorithm, such as the number k of clusters. In spite of the popularity
of stability in practical applications, there has been very little theoretical
analysis of this notion. In this paper we provide a formal definition
of stability and analyze some of its basic properties. Quite surprisingly,
the conclusion of our analysis is that for large sample size, stability is
fully determined by the behavior of the objective function which the
clustering algorithm is aiming to minimize. If the objective function has
a unique global minimizer, the algorithm is stable, otherwise it is unstable.
In particular we conclude that stability is not a well-suited tool
to determine the number of clusters - it is determined by the symmetries
of the data which may be unrelated to clustering parameters. We
prove our results for center-based clusterings and for spectral clustering,
and support our conclusions by many examples in which the behavior of
stability is counter-intuitive.