We propose an independence criterion based on the eigenspectrum of covariance operators in reproducing kernel Hilbert spaces (RKHSs), consisting of an empirical estimate of the Hilbert-Schmidt norm of the cross-covariance operator (we term this a Hilbert-Schmidt Independence Criterion, or HSIC). This approach has several advantages, compared with previous kernel-based independence criteria. First, the empirical estimate is simpler than any other kernel dependence test, and requires no user-defined regularisation. Second, there is a clearly defined population quantity which the empirical estimate approaches in the large sample limit, with exponential convergence guaranteed between the two: this ensures that independence tests based on HSIC do not suffer from slow learning rates. Finally, we show in the context of independent component analysis (ICA) that the performance of HSIC is competitive with that of previously published kernel-based criteria, and of other recently published ICA methods.
We propose new bounds on the error of learning algorithms in terms of a data-dependent notion of complexity. The estimates we establish give optimal rates and are based on a local and empirical version of Rademacher averages, in the sense that the Rademacher averages are computed from the data, on a subset of functions with small empirical error. We present some applications to classification and prediction with convex function classes, and with kernel classes in particular.
Consistency is a key property of all statistical procedures analyzing
randomly sampled data. Surprisingly, despite decades of work, little is known
about consistency of most clustering algorithms. In this paper we investigate
consistency of the popular family of spectral clustering algorithms, which
clusters the data with the help of eigenvectors of graph Laplacian matrices. We
develop new methods to establish that, for increasing sample size, those
eigenvectors converge to the eigenvectors of certain limit operators. As a
result, we can prove that one of the two major classes of spectral clustering
(normalized clustering) converges under very general conditions, while the
other (unnormalized clustering) is only consistent under strong additional
assumptions, which are not always satisfied in real data. We conclude that our
analysis provides strong evidence for the superiority of normalized spectral
clustering.Comment: Published in at http://dx.doi.org/10.1214/009053607000000640 the
Annals of Statistics (http://www.imstat.org/aos/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Abstract. The last few years have witnessed important new developments in the theory and practice of pattern classification. We intend to survey some of the main new ideas that have led to these recent results.Mathematics Subject Classification. 62G08, 60E15, 68Q32.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.