Goto

Collaborating Authors

 Semantic Networks


Modeling the Diachronic Evolution of Legal Norms: An LRMoo-Based, Component-Level, Event-Centric Approach to Legal Knowledge Graphs

de Martim, Hudson

arXiv.org Artificial Intelligence

Representing the temporal evolution of legal norms is a critical challenge for automated processing. While foundational frameworks exist, they lack a formal pattern for granular, component-level versioning, hindering the deterministic point-in-time reconstruction of legal texts required by reliable AI applications. This paper proposes a structured, temporal modeling pattern grounded in the LRMoo ontology. Our approach models a norm's evolution as a diachronic chain of versioned F1 Works, distinguishing between language-agnostic Temporal Versions (TV)-each being a distinct Work-and their monolingual Language Versions (LV), modeled as F2 Expressions. The legislative amendment process is formalized through event-centric modeling, allowing changes to be traced precisely. Using the Brazilian Constitution as a case study, we demonstrate that our architecture enables the exact reconstruction of any part of a legal text as it existed on a specific date. This provides a verifiable semantic backbone for legal knowledge graphs, offering a deterministic foundation for trustworthy legal AI.


HyperComplEx: Adaptive Multi-Space Knowledge Graph Embeddings

Gajjar, Jugal, Ranaware, Kaustik, Subramaniakuppusamy, Kamalasankari, Gandhi, Vaibhav

arXiv.org Artificial Intelligence

Knowledge graphs have emerged as fundamental structures for representing complex relational data across scientific and enterprise domains. However, existing embedding methods face critical limitations when modeling diverse relationship types at scale: Euclidean models struggle with hierarchies, vector space models cannot capture asymmetry, and hyperbolic models fail on symmetric relations. We propose HyperComplEx, a hybrid embedding framework that adaptively combines hyperbolic, complex, and Euclidean spaces via learned attention mechanisms. A relation-specific space weighting strategy dynamically selects optimal geometries for each relation type, while a multi-space consistency loss ensures coherent predictions across spaces. We evaluate HyperComplEx on computer science research knowledge graphs ranging from 1K papers (~25K triples) to 10M papers (~45M triples), demonstrating consistent improvements over state-of-the-art baselines including TransE, RotatE, DistMult, ComplEx, SEPA, and UltraE. Additional tests on standard benchmarks confirm significantly higher results than all baselines. On the 10M-paper dataset, HyperComplEx achieves 0.612 MRR, a 4.8% relative gain over the best baseline, while maintaining efficient training, achieving 85 ms inference per triple. The model scales near-linearly with graph size through adaptive dimension allocation. We release our implementation and dataset family to facilitate reproducible research in scalable knowledge graph embeddings.


AI Agent-Driven Framework for Automated Product Knowledge Graph Construction in E-Commerce

Peshevski, Dimitar, Stojanov, Riste, Trajanov, Dimitar

arXiv.org Artificial Intelligence

The rapid expansion of e-commerce platforms generates vast amounts of unstructured product data, creating significant challenges for information retrieval, recommendation systems, and data analytics. Knowledge Graphs (KGs) offer a structured, interpretable format to organize such data, yet constructing product-specific KGs remains a complex and manual process. This paper introduces a fully automated, AI agent-driven framework for constructing product knowledge graphs directly from unstructured product descriptions. Leveraging Large Language Models (LLMs), our method operates in three stages using dedicated agents: ontology creation and expansion, ontology refinement, and knowledge graph population. This agent-based approach ensures semantic coherence, scalability, and high-quality output without relying on predefined schemas or handcrafted extraction rules. We evaluate the system on a real-world dataset of air conditioner product descriptions, demonstrating strong performance in both ontology generation and KG population. The framework achieves over 97\% property coverage and minimal redundancy, validating its effectiveness and practical applicability. Our work highlights the potential of LLMs to automate structured knowledge extraction in retail, providing a scalable path toward intelligent product data integration and utilization.


Improving Continual Learning of Knowledge Graph Embeddings via Informed Initialization

Pons, Gerard, Bilalli, Besim, Queralt, Anna

arXiv.org Artificial Intelligence

Many Knowledege Graphs (KGs) are frequently updated, forcing their Knowledge Graph Embeddings (KGEs) to adapt to these changes. To address this problem, continual learning techniques for KGEs incorporate embeddings for new entities while updating the old ones. One necessary step in these methods is the initialization of the embeddings, as an input to the KGE learning process, which can have an important impact in the accuracy of the final embeddings, as well as in the time required to train them. This is especially relevant for relatively small and frequent updates. We propose a novel informed embedding initialization strategy, which can be seamlessly integrated into existing continual learning methods for KGE, that enhances the acquisition of new knowledge while reducing catastrophic forgetting. Specifically, the KG schema and the previously learned embeddings are utilized to obtain initial representations for the new entities, based on the classes the entities belong to. Our extensive experimental analysis shows that the proposed initialization strategy improves the predictive performance of the resulting KGEs, while also enhancing knowledge retention. Furthermore, our approach accelerates knowledge acquisition, reducing the number of epochs, and therefore time, required to incrementally learn new embeddings. Finally, its benefits across various types of KGE learning models are demonstrated.



Supplementary Material of Rot-Pro: Modeling Transitivity by Projection in Knowledge Graph Embedding

Neural Information Processing Systems

In section 3.2 of the submitted paper, we use the conclusion that "the transitive relation can be represented as the union of transitive closures of of all transitive chains." S1, S2, and S3 datasets of Counties are separated by '/'. Our model is implemented in Python 3.6 using Pytorch 1.1.0. We list the best hyper-parameter setting of Rot-Pro on the above datasets in Table 2. The fully expressive of BoxE refers to that it is able to express inference patterns, which includes symmetry, anti-symmetry, inversion, composition, hierarchy, intersection, and mutual exclusion.


Rot-Pro: Modeling Transitivity by Projection in Knowledge Graph Embedding

Neural Information Processing Systems

Knowledge graph embedding models learn the representations of entities and relations in the knowledge graphs for predicting missing links (relations) between entities.


Duality-Induced Regularizer for Tensor Factorization Based Knowledge Graph Completion Supplementary Material

Neural Information Processing Systems

Theorem 1. Suppose that ˆ X In DB models, the commonly used p is either 1 or 2. When p = 2, DURA takes the form as the one in Equation (8) in the main text. If p = 1, we cannot expand the squared score function of the associated DB models as in Equation (4). Therefore, we choose p = 2 . 2 Table 2: Hyperparameters found by grid search. Suppose that k is the number of triplets known to be true in the knowledge graph, n is the embedding dimension of entities. That is to say, the computational complexity of weighted DURA is the same as the weighted squared Frobenius norm regularizer.



A Theory of Link Prediction via Relational Weisfeiler-Leman on Knowledge Graphs

Neural Information Processing Systems

Our goal is to provide a systematic understanding of the landscape of graph neural networks for knowledge graphs pertaining to the prominent task of link prediction. Our analysis entails a unifying perspective on seemingly unrelated models and unlocks a series of other models.