Hara, Kazuo
Flattening the Density Gradient for Eliminating Spatial Centrality to Reduce Hubness
Hara, Kazuo (National Institute of Genetics) | Suzuki, Ikumi (Yamagata University) | Kobayashi, Kei (The Institute of Statistical Mathematics) | Fukumizu, Kenji (The Institute of Statistical Mathematics) | Radovanovic, Milos (University of Novi Sad)
Spatial centrality, whereby samples closer to the center of a dataset tend to be closer to all other samples, is regarded as one source of hubness. Hubness is well known to degrade k-nearest-neighbor (k-NN) classification. Spatial centrality can be removed by centering, i.e., shifting the origin to the global center of the dataset, in cases where inner product similarity is used. However, when Euclidean distance is used, centering has no effect on spatial centrality because the distance between the samples is the same before and after centering. As described in this paper, we propose a solution for the hubness problem when Euclidean distance is considered. We provide a theoretical explanation to demonstrate how the solution eliminates spatial centrality and reduces hubness. We then present some discussion of the reason the proposed solution works, from a viewpoint of density gradient, which is regarded as the origin of spatial centrality and hubness. We demonstrate that the solution corresponds to flattening the density gradient. Using real-world datasets, we demonstrate that the proposed method improves k-NN classification performance and outperforms an existing hub-reduction method.
Ridge Regression, Hubness, and Zero-Shot Learning
Shigeto, Yutaro, Suzuki, Ikumi, Hara, Kazuo, Shimbo, Masashi, Matsumoto, Yuji
This paper discusses the effect of hubness in zero-shot learning, when ridge regression is used to find a mapping between the example space to the label space. Contrary to the existing approach, which attempts to find a mapping from the example space to the label space, we show that mapping labels into the example space is desirable to suppress the emergence of hubs in the subsequent nearest neighbor search step. Assuming a simple data model, we prove that the proposed approach indeed reduces hubness. This was verified empirically on the tasks of bilingual lexicon extraction and image labeling: hubness was reduced with both of these tasks and the accuracy was improved accordingly.
Localized Centering: Reducing Hubness in Large-Sample Data
Hara, Kazuo (National Institute of Genetics) | Suzuki, Ikumi (National Institute of Genetics) | Shimbo, Masashi (Nara Institute of Science and Technology) | Kobayashi, Kei (The Institute of Statistical Mathematics) | Fukumizu, Kenji (The Institute of Statistical Mathematics) | Radovanović, Miloš (University of Novi Sad)
Hubness has been recently identified as a problematic phenomenon occurring in high-dimensional space. In this paper, we address a different type of hubness that occurs when the number of samples is large. We investigate the difference between the hubness in high-dimensional data and the one in large-sample data. One finding is that centering, which is known to reduce the former, does not work for the latter. We then propose a new hub-reduction method, called localized centering. It is an extension of centering, yet works effectively for both types of hubness. Using real-world datasets consisting of a large number of documents, we demonstrate that the proposed method improves the accuracy of k-nearest neighbor classification.
Investigating the Effectiveness of Laplacian-Based Kernels in Hub Reduction
Suzuki, Ikumi (Nara Institute of Science and Technology) | Hara, Kazuo (National Institute of Genetics) | Shimbo, Masashi (Nara Institute of Science and Technology) | Matsumoto, Yuji (Nara Institute of Science and Technology) | Saerens, Marco (Universite Catholique de Louvain)
A “hub” is an object closely surrounded by, or very similar to, many other objects in the dataset. Recent studies by Radovanovi´c et al. indicate that in high dimensional spaces, hubs almost always emerge, and objects close to the data centroid tend to become hubs. In this paper, we show that the family of kernels based on the graph Laplacian makes all objects in the dataset equally similar to the centroid, and thus they are expected to make less hubs when used as a similarity measure. We investigate this hypothesis using both synthetic and real-world data. It turns out that these kernels suppress hubs in some cases but not always, and the results seem to be affected by the size of the data—a factor not discussed previously. However, for the datasets in which hubs are indeed reduced by the Laplacian-based kernels, these kernels work well in ranking and classification tasks. This result suggests that the amount of hubs, which can be readily computed in an unsupervised fashion, can be a yardstick of whether Laplacian-based kernels work effectively for a given data.