Nguyen, Minh Le
A Decoupling and Aggregating Framework for Joint Extraction of Entities and Relations
Wang, Yao, Liu, Xin, Kong, Weikun, Yu, Hai-Tao, Racharak, Teeradaj, Kim, Kyoung-Sook, Nguyen, Minh Le
Named Entity Recognition and Relation Extraction are two crucial and challenging subtasks in the field of Information Extraction. Despite the successes achieved by the traditional approaches, fundamental research questions remain open. First, most recent studies use parameter sharing for a single subtask or shared features for both two subtasks, ignoring their semantic differences. Second, information interaction mainly focuses on the two subtasks, leaving the fine-grained informtion interaction among the subtask-specific features of encoding subjects, relations, and objects unexplored. Motivated by the aforementioned limitations, we propose a novel model to jointly extract entities and relations. The main novelties are as follows: (1) We propose to decouple the feature encoding process into three parts, namely encoding subjects, encoding objects, and encoding relations. Thanks to this, we are able to use fine-grained subtask-specific features. The experimental results demonstrate that our model outperforms several previous state-of-the-art models. Extensive additional experiments further confirm the effectiveness of our model. A Decoupling and Aggregating Framework for Joint Extraction of Entities and Relations Introduction Named Entity Recognition (NER) and Relation Extraction (RE), as two essential subtasks in information extraction, aim to extract entities and relations from semi-structured and unstructured texts. They are used in many downstream applications in different domains, such as knowledge graph construction [38, 39], Question-Answering [36, 37], and knowledge graph-based recommendation system [40, 41]. Most traditional models and some methods used in specialized areas [9,35,43,46] construct separate models for NER and RE to extract entities and relations in a pipelined manner. This type of method suffers from error propagation and unilateral information interaction.
Encoded Summarization: Summarizing Documents into Continuous Vector Space for Legal Case Retrieval
Tran, Vu, Nguyen, Minh Le, Tojo, Satoshi, Satoh, Ken
On the other hand, we explore the benefits from combining lexical features and latent features generated with neural networks. Our experiments show that lexical features and latent features generated with neural networks complement each other to improve the retrieval system performance. Furthermore, our experimental results suggest the importance of case summarization in different aspects: using provided summaries and performing encoded summarization. Our approach achieved F1 of 65.6% and 57.6% on the experimental datasets of legal case retrieval tasks.
Causal Intersectionality and Dual Form of Gradient Descent for Multimodal Analysis: a Case Study on Hateful Memes
Miyanishi, Yosuke, Nguyen, Minh Le
In the wake of the explosive growth of machine learning (ML) usage, particularly within the context of emerging Large Language Models (LLMs), comprehending the semantic significance rooted in their internal workings is crucial. While causal analyses focus on defining semantics and its quantification, the gradient-based approach is central to explainable AI (XAI), tackling the interpretation of the black box. By synergizing these approaches, the exploration of how a model's internal mechanisms illuminate its causal effect has become integral for evidence-based decision-making. A parallel line of research has revealed that intersectionality - the combinatory impact of multiple demographics of an individual - can be structured in the form of an Averaged Treatment Effect (ATE). Initially, this study illustrates that the hateful memes detection problem can be formulated as an ATE, assisted by the principles of intersectionality, and that a modality-wise summarization of gradient-based attention attribution scores can delineate the distinct behaviors of three Transformerbased models concerning ATE. Subsequently, we show that the latest LLM LLaMA2 has the ability to disentangle the intersectional nature of memes detection in an in-context learning setting, with their mechanistic properties elucidated via meta-gradient, a secondary form of gradient. In conclusion, this research contributes to the ongoing dialogue surrounding XAI and the multifaceted nature of ML models.
Miko Team: Deep Learning Approach for Legal Question Answering in ALQAC 2022
Van, Hieu Nguyen, Nguyen, Dat, Nguyen, Phuong Minh, Nguyen, Minh Le
We introduce efficient deep learning-based methods for legal document processing including Legal Document Retrieval and Legal Question Answering tasks in the Automated Legal Question Answering Competition (ALQAC 2022). In this competition, we achieve 1\textsuperscript{st} place in the first task and 3\textsuperscript{rd} place in the second task. Our method is based on the XLM-RoBERTa model that is pre-trained from a large amount of unlabeled corpus before fine-tuning to the specific tasks. The experimental results showed that our method works well in legal retrieval information tasks with limited labeled data. Besides, this method can be applied to other information retrieval tasks in low-resource languages.
ParaLaw Nets -- Cross-lingual Sentence-level Pretraining for Legal Text Processing
Nguyen, Ha-Thanh, Tran, Vu, Nguyen, Phuong Minh, Vuong, Thi-Hai-Yen, Bui, Quan Minh, Nguyen, Chau Minh, Dang, Binh Tran, Nguyen, Minh Le, Satoh, Ken
Ambiguity is a characteristic of natural language, which makes expression ideas flexible. However, in a domain that requires accurate statements, it becomes a barrier. Specifically, a single word can have many meanings and multiple words can have the same meaning. When translating a text into a foreign language, the translator needs to determine the exact meaning of each element in the original sentence to produce the correct translation sentence. From that observation, in this paper, we propose ParaLaw Nets, a pretrained model family using sentence-level cross-lingual information to reduce ambiguity and increase the performance in legal text processing. This approach achieved the best result in the Question Answering task of COLIEE-2021.