Dubey, Mohnish
Semantic Answer Type and Relation Prediction Task (SMART 2021)
Mihindukulasooriya, Nandana, Dubey, Mohnish, Gliozzo, Alfio, Lehmann, Jens, Ngomo, Axel-Cyrille Ngonga, Usbeck, Ricardo, Rossiello, Gaetano, Kumar, Uttam
Each year the International Semantic Web Conference organizes a set of Semantic Web Challenges to establish competitions that will advance state-of-the-art solutions in some problem domains. The Semantic Answer Type and Relation Prediction Task (SMART) task is one of the ISWC 2021 Semantic Web challenges. This is the second year of the challenge after a successful SMART 2020 at ISWC 2020. This year's version focuses on two sub-tasks that are very important to Knowledge Base Question Answering (KBQA): Answer Type Prediction and Relation Prediction. Question type and answer type prediction can play a key role in knowledge base question answering systems providing insights about the expected answer that are helpful to generate correct queries or rank the answer candidates. More concretely, given a question in natural language, the first task is, to predict the answer type using a target ontology (e.g., DBpedia or Wikidata. Similarly, the second task is to identify relations in the natural language query and link them to the relations in a target ontology. This paper discusses the task descriptions, benchmark datasets, and evaluation metrics. For more information, please visit https://smart-task.github.io/2021/.
SeMantic AnsweR Type prediction task (SMART) at ISWC 2020 Semantic Web Challenge
Mihindukulasooriya, Nandana, Dubey, Mohnish, Gliozzo, Alfio, Lehmann, Jens, Ngomo, Axel-Cyrille Ngonga, Usbeck, Ricardo
Question Answering (QA) is a popular task in Natural Language Processing and Information Retrieval, in which the goal is to answer a natural language question (going beyond the document retrieval). There are further sub-tasks, for instance, reading comprehension, in which the expected answers can be either a segment of text or span, from the corresponding reading passage of text. The Stanford Question Answering Dataset (SQuAD) [Rajpurkar et al., 2018] is an example of this task. Similarly, another task is Question Answering over Knowledge Bases, in which the expected answer can either be a set of entities in the knowledge base or an answer derived from an aggregation of them. Question Answering over Linked Data (QALD) [Usbeck et al., 2018] and Large Scale Complex Question Answering Dataset (LC-QuAD) [Dubey et al., 2019] are two examples for this task Question or answer type classification plays a key role in question answering [Harabagiu et al., 2000, Allam and Haggag, 2012].
PNEL: Pointer Network based End-To-End Entity Linking over Knowledge Graphs
Banerjee, Debayan, Chaudhuri, Debanjan, Dubey, Mohnish, Lehmann, Jens
Question Answering systems are generally modelled as a pipeline consisting of a sequence of steps. In such a pipeline, Entity Linking (EL) is often the first step. Several EL models first perform span detection and then entity disambiguation. In such models errors from the span detection phase cascade to later steps and result in a drop of overall accuracy. Moreover, lack of gold entity spans in training data is a limiting factor for span detector training. Hence the movement towards end-to-end EL models began where no separate span detection step is involved. In this work we present a novel approach to end-to-end EL by applying the popular Pointer Network model, which achieves competitive performance. We demonstrate this in our evaluation over three datasets on the Wikidata Knowledge Graph.
EARL: Joint Entity and Relation Linking for Question Answering over Knowledge Graphs
Dubey, Mohnish, Banerjee, Debayan, Chaudhuri, Debanjan, Lehmann, Jens
Many question answering systems over knowledge graphs rely on entity and relation linking components in order to connect the natural language input to the underlying knowledge graph. Traditionally, entity linking and relation linking have been performed either as dependent sequential tasks or as independent parallel tasks. In this paper, we propose a framework called EARL, which performs entity linking and relation linking as a joint task. EARL implements two different solution strategies for which we provide a comparative analysis in this paper: The first strategy is a formalisation of the joint entity and relation linking tasks as an instance of the Generalised Travelling Salesman Problem (GTSP). In order to be computationally feasible, we employ approximate GTSP solvers. The second strategy uses machine learning in order to exploit the connection density between nodes in the knowledge graph. It relies on three base features and re-ranking steps in order to predict entities and relations. We compare the strategies and evaluate them on a dataset with 5000 questions. Both strategies significantly outperform the current state-of-the-art approaches for entity and relation linking.