Mihaylova, Tsvetomila
Do Visual-Language Maps Capture Latent Semantics?
Pekkanen, Matti, Mihaylova, Tsvetomila, Verdoja, Francesco, Kyrki, Ville
Visual-language models (VLMs) have recently been introduced in robotic mapping by using the latent representations, i.e., embeddings, of the VLMs to represent the natural language semantics in the map. The main benefit is moving beyond a small set of human-created labels toward open-vocabulary scene understanding. While there is anecdotal evidence that maps built this way support downstream tasks, such as navigation, rigorous analysis of the quality of the maps using these embeddings is lacking. We investigate two critical properties of map quality: queryability and consistency. The evaluation of queryability addresses the ability to retrieve information from the embeddings. We investigate two aspects of consistency: intra-map consistency and inter-map consistency. Intra-map consistency captures the ability of the embeddings to represent abstract semantic classes, and inter-map consistency captures the generalization properties of the representation. In this paper, we propose a way to analyze the quality of maps created using VLMs, which forms an open-source benchmark to be used when proposing new open-vocabulary map representations. We demonstrate the benchmark by evaluating the maps created by two state-of-the-art methods, VLMaps and OpenScene, using two encoders, LSeg and OpenSeg, using real-world data from the Matterport3D data set. We find that OpenScene outperforms VLMaps with both encoders, and LSeg outperforms OpenSeg with both methods.
Discrete Latent Structure in Neural Networks
Niculae, Vlad, Corro, Caio F., Nangia, Nikita, Mihaylova, Tsvetomila, Martins, André F. T.
Many types of data from fields including natural language processing, computer vision, and bioinformatics, are well represented by discrete, compositional structures such as trees, sequences, or matchings. Latent structure models are a powerful tool for learning to extract such representations, offering a way to incorporate structural bias, discover insight about the data, and interpret decisions. However, effective training is challenging, as neural networks are typically designed for continuous computation. This text explores three broad strategies for learning with discrete latent structure: continuous relaxation, surrogate gradients, and probabilistic estimation. Our presentation relies on consistent notations for a wide range of models. As such, we reveal many new connections between latent structure learning strategies, showing how most consist of the same small set of fundamental building blocks, but use them differently, leading to substantially different applicability and properties.
SUper Team at SemEval-2016 Task 3: Building a feature-rich system for community question answering
Mihaylova, Tsvetomila, Gencheva, Pepa, Boyanov, Martin, Yovcheva, Ivana, Mihaylov, Todor, Hardalov, Momchil, Kiprov, Yasen, Balchev, Daniel, Koychev, Ivan, Nakov, Preslav, Nikolova, Ivelina, Angelova, Galia
We present the system we built for participating in SemEval-2016 Task 3 on Community Question Answering. We achieved the best results on subtask C, and strong results on subtasks A and B, by combining a rich set of various types of features: semantic, lexical, metadata, and user-related. The most important group turned out to be the metadata for the question and for the comment, semantic vectors trained on QatarLiving data and similarities between the question and the comment for subtasks A and C, and between the original and the related question for Subtask B.
Automatic Fact-Checking Using Context and Discourse Information
Atanasova, Pepa, Nakov, Preslav, Màrquez, Lluís, Barrón-Cedeño, Alberto, Karadzhov, Georgi, Mihaylova, Tsvetomila, Mohtarami, Mitra, Glass, James
We study the problem of automatic fact-checking, paying special attention to the impact of contextual and discourse information. We address two related tasks: (i) detecting check-worthy claims, and (ii) fact-checking claims. We develop supervised systems based on neural networks, kernel-based support vector machines, and combinations thereof, which make use of rich input representations in terms of discourse cues and contextual features. For the check-worthiness estimation task, we focus on political debates, and we model the target claim in the context of the full intervention of a participant and the previous and the following turns in the debate, taking into account contextual meta information. For the fact-checking task, we focus on answer verification in a community forum, and we model the veracity of the answer with respect to the entire question--answer thread in which it occurs as well as with respect to other related posts from the entire forum. We develop annotated datasets for both tasks and we run extensive experimental evaluation, confirming that both types of information ---but especially contextual features--- play an important role.
SemEval-2019 Task 8: Fact Checking in Community Question Answering Forums
Mihaylova, Tsvetomila, Karadjov, Georgi, Atanasova, Pepa, Baly, Ramy, Mohtarami, Mitra, Nakov, Preslav
We present SemEval-2019 Task 8 on Fact Checking in Community Question Answering Forums, which features two subtasks. Subtask A is about deciding whether a question asks for factual information vs. an opinion/advice vs. just socializing. Subtask B asks to predict whether an answer to a factual question is true, false or not a proper answer. We received 17 official submissions for subtask A and 11 official submissions for Subtask B. For subtask A, all systems improved over the majority class baseline. For Subtask B, all systems were below a majority class baseline, but several systems were very close to it. The leaderboard and the data from the competition can be found at http://competitions.codalab.org/competitions/20022
Fact Checking in Community Forums
Mihaylova, Tsvetomila (Sofia University "St. Kliment Ohridski") | Nakov, Preslav ( Qatar Computing Research Institute, HBKU ) | Màrquez, Lluís (Qatar Computing Research Institute, HBKU) | Barrón-Cedeño, Alberto (Qatar Computing Research Institute, HBKU) | Mohtarami, Mitra (Massachusetts Institute of Technology) | Karadzhov, Georgi (Sofia University "St. Kliment Ohridski") | Glass, James (Massachusetts Institute of Technology)
Community Question Answering (cQA) forums are very popular nowadays, as they represent effective means for communities around particular topics to share information. Unfortunately, this information is not always factual. Thus, here we explore a new dimension in the context of cQA, which has been ignored so far: checking the veracity of answers to particular questions in cQA forums. As this is a new problem, we create a specialized dataset for it. We further propose a novel multi-faceted model, which captures information from the answer content (what is said and how), from the author profile (who says it), from the rest of the community forum (where it is said), and from external authoritative sources of information (external support). Evaluation results show a MAP value of 86.54, which is 21 points absolute above the baseline.