mncut
ASimpleApproachtoAutomatedSpectralClustering Appendices
Let ˆc be the optimal solution of minimizec 12kφ(y) φ(X)ck2 + λ2kck2, where φ is induced by Gaussian kernel and y is arbitrary. It is worth noting that Algorithm 1 can be easily implemented parallelly, which will reduce the time complexity to O(max(m,r)n2 +kmn). Denote vi = (vi1,...,vin) the i-th row of V and let vi = (vi1,...,vid), where d < n. Clustering the columns of X given by Definition C.1 according to the polynomials is actually a manifold clustering problem beyond the setting of subspaceclustering. The following theorem verifies the effectiveness of (15) followed by the truncation operation in manifolddetection. 2 TheoremC.3.
- Asia > China > Guangdong Province > Shenzhen (0.05)
- Asia > China > Beijing > Beijing (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (3 more...)
EGGS: Eigen-Gap Guided Search Making Subspace Clustering Easy
Fan, Jicong, Tu, Yiheng, Zhang, Zhao, Zhao, Mingbo
The performance of spectral clustering heavily relies on the quality of affinity matrix. A variety of affinity-matrix-construction methods have been proposed but they have hyper-parameters to determine beforehand, which requires strong experience and lead to difficulty in real applications especially when the inter-cluster similarity is high or/and the dataset is large. On the other hand, we often have to determine to use a linear model or a nonlinear model, which still depends on experience. To solve these two problems, in this paper, we present an eigen-gap guided search method for subspace clustering. The main idea is to find the most reliable affinity matrix among a set of candidates constructed by linear and kernel regressions, where the reliability is quantified by the \textit{relative-eigen-gap} of graph Laplacian defined in this paper. We show, theoretically and numerically, that the Laplacian matrix with a larger relative-eigen-gap often yields a higher clustering accuracy and stability. Our method is able to automatically search the best model and hyper-parameters in a pre-defined space. The search space is very easy to determine and can be arbitrarily large, though a relatively compact search space can reduce the highly unnecessary computation. Our method has high flexibility and convenience in real applications, and also has low computational cost because the affinity matrix is not computed by iterative optimization. We extend the method to large-scale datasets such as MNIST, on which the time cost is less than 90s and the clustering accuracy is state-of-the-art. Extensive experiments of natural image clustering show that our method is more stable, accurate, and efficient than baseline methods.
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (0.75)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Clustering (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Unsupervised spectral learning
Shortreed, Susan, Meila, Marina
In spectral clustering and spectral image segmentation, the data is partioned starting from a given matrix of pairwise similarities S. the matrix S is constructed by hand, or learned on a separate training set. In this paper we show how to achieve spectral clustering in unsupervised mode. Our algorithm starts with a set of observed pairwise features, which are possible components of an unknown, parametric similarity function. This function is learned iteratively, at the same time as the clustering of the data. The algorithm shows promosing results on synthetic and real data.
- North America > United States > Washington > King County > Seattle (0.14)
- Asia > Middle East > Jordan (0.05)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.05)
- North America > United States > California > Alameda County > Berkeley (0.04)