Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces
Nayyeri, Mojtaba, Wang, Zihao, Akter, Mst. Mahfuja, Alam, Mirza Mohtashim, Rony, Md Rashad Al Hasan, Lehmann, Jens, Staab, Steffen
–arXiv.org Artificial Intelligence
Knowledge graphs comprise structural and textual information to represent knowledge. To predict new structural knowledge, current approaches learn representations using both types of information through knowledge graph embeddings and language models. These approaches commit to a single pre-trained language model. We hypothesize that heterogeneous language models may provide complementary information not exploited by current approaches. To investigate this hypothesis, we propose a unified framework that integrates multiple representations of structural knowledge and textual information. Our approach leverages hypercomplex algebra to model the interactions between (i) graph structural information and (ii) multiple text representations. Specifically, we utilize Dihedron models with 4*D dimensional hypercomplex numbers to integrate four different representations: structural knowledge graph embeddings, word-level representations (e.g., Word2vec and Fast-Text), sentence-level representations (using a sentence transformer), and document-level representations (using FastText or Doc2vec). Our unified framework score the plausibility of labeled edges via Dihedron products, thus modeling pairwise interactions between the four representations. Extensive experimental evaluations on standard benchmark datasets confirm our hypothesis showing the superiority of our two new frameworks for link prediction tasks.
arXiv.org Artificial Intelligence
Aug-16-2023
- Country:
- Europe > Germany (0.68)
- North America > United States (0.46)
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Leisure & Entertainment (1.00)
- Technology: