Goto

Collaborating Authors

 ica-transformed


Transforming Hidden States into Binary Semantic Features

Musil, Tomáš, Mareček, David

arXiv.org Artificial Intelligence

However, with 2. centering the data (setting the mean to zero) the advance of Large Language Models (LLMs), and whitening them (setting variance of each this inspiration has become rather indirect. In this component to 1), paper, we show that distributional theories of meaning can still be relevant in interpreting the hidden 3. iteratively finding directions in the data that states of LLMs and that Independent Component are the most non-Gaussian. Analysis (ICA) can help us overcome some of The last step is based on the assumption of the the challenges associated with understanding these central limit theorem: the mixed signal is a sum complex models.


Revisiting Cosine Similarity via Normalized ICA-transformed Embeddings

Yamagiwa, Hiroaki, Oyama, Momose, Shimodaira, Hidetoshi

arXiv.org Artificial Intelligence

Cosine similarity is widely used to measure the similarity between two embeddings, while interpretations based on angle and correlation coefficient are common. In this study, we focus on the interpretable axes of embeddings transformed by Independent Component Analysis (ICA), and propose a novel interpretation of cosine similarity as the sum of semantic similarities over axes. To investigate this, we first show experimentally that unnormalized embeddings contain norm-derived artifacts. We then demonstrate that normalized ICA-transformed embeddings exhibit sparsity, with a few large values in each axis and across embeddings, thereby enhancing interpretability by delineating clear semantic contributions. Finally, to validate our interpretation, we perform retrieval experiments using ideal embeddings with and without specific semantic components.


Axis Tour: Word Tour Determines the Order of Axes in ICA-transformed Embeddings

Yamagiwa, Hiroaki, Takase, Yusuke, Shimodaira, Hidetoshi

arXiv.org Artificial Intelligence

Word embedding is one of the most important components in natural language processing, but interpreting high-dimensional embeddings remains a challenging problem. To address this problem, Independent Component Analysis (ICA) is identified as an effective solution. ICA-transformed word embeddings reveal interpretable semantic axes; however, the order of these axes are arbitrary. In this study, we focus on this property and propose a novel method, Axis Tour, which optimizes the order of the axes. Inspired by Word Tour, a onedimensional word embedding method, we aim to improve the clarity of the word embedding space by maximizing the semantic continuity of the axes. Furthermore, we show through experiments Figure 1: Scatterplots of normalized ICA-transformed on downstream tasks that Axis Tour word embeddings whose axes are ordered by Axis Tour constructs better low-dimensional embeddings and Skewness Sort. In the upper part, Axis Tour is applied compared to both PCA and ICA.


Discovering Universal Geometry in Embeddings with ICA

Yamagiwa, Hiroaki, Oyama, Momose, Shimodaira, Hidetoshi

arXiv.org Artificial Intelligence

This study utilizes Independent Component Analysis (ICA) to unveil a consistent semantic structure within embeddings of words or images. Our approach extracts independent semantic components from the embeddings of a pre-trained model by leveraging anisotropic information that remains after the whitening process in Principal Component Analysis (PCA). We demonstrate that each embedding can be expressed as a composition of a few intrinsic interpretable axes and that these semantic axes remain consistent across different languages, algorithms, and modalities. The discovery of a universal semantic structure in the geometric patterns of embeddings enhances our understanding of the representations in embeddings.