Abstract
The
k
-means method is one of the most widely used clustering algorithms, drawing its popularity from its speed in practice. Recently, however, it was shown to have exponential worst-case running time. In order to close the gap between practical performance and theoretical analysis, the
k
-means method has been studied in the model of smoothed analysis. But even the smoothed analyses so far are unsatisfactory as the bounds are still super-polynomial in the number
n
of data points.
In this article, we settle the smoothed running time of the
k
-means method. We show that the smoothed number of iterations is bounded by a polynomial in
n
and 1/
σ
, where
σ
is the standard deviation of the Gaussian perturbations. This means that if an arbitrary input data set is randomly perturbed, then the
k
-means method will run in expected polynomial time on that input set.
Publisher
Association for Computing Machinery (ACM)
Subject
Artificial Intelligence,Hardware and Architecture,Information Systems,Control and Systems Engineering,Software
Cited by
49 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献