Goto

Collaborating Authors

 Šuppa, Marek


MMTEB: Massive Multilingual Text Embedding Benchmark

arXiv.org Artificial Intelligence

Text embeddings are typically evaluated on a limited set of tasks, which are constrained by language, domain, and task diversity. To address these limitations and provide a more comprehensive evaluation, we introduce the Massive Multilingual Text Embedding Benchmark (MMTEB) - a large-scale, community-driven expansion of MTEB, covering over 500 quality-controlled evaluation tasks across 250+ languages. MMTEB includes a diverse set of challenging, novel tasks such as instruction following, long-document retrieval, and code retrieval, representing the largest multilingual collection of evaluation tasks for embedding models to date. Using this collection, we develop several highly multilingual benchmarks, which we use to evaluate a representative set of models. We find that while large language models (LLMs) with billions of parameters can achieve state-of-the-art performance on certain language subsets and task categories, the best-performing publicly available model is multilingual-e5-large-instruct with only 560 million parameters. To facilitate accessibility and reduce computational cost, we introduce a novel downsampling method based on inter-task correlation, ensuring a diverse selection while preserving relative model rankings. Furthermore, we optimize tasks such as retrieval by sampling hard negatives, creating smaller but effective splits. These optimizations allow us to introduce benchmarks that drastically reduce computational demands. For instance, our newly introduced zero-shot English benchmark maintains a ranking order similar to the full-scale version but at a fraction of the computational cost.


RoBo6: Standardized MMT Light Curve Dataset for Rocket Body Classification

arXiv.org Artificial Intelligence

Space debris presents a critical challenge for the sustainability of future space missions, emphasizing the need for robust and standardized identification methods. However, a comprehensive benchmark for rocket body classification remains absent. This paper addresses this gap by introducing the RoBo6 dataset for rocket body classification based on light curves. The dataset, derived from the Mini Mega Tortora database, includes light curves for six rocket body classes: CZ-3B, Atlas 5 Centaur, Falcon 9, H-2A, Ariane 5, and Delta 4. With 5,676 training and 1,404 test samples, it addresses data inconsistencies using resampling, normalization, and filtering techniques. Several machine learning models were evaluated, including CNN and transformer-based approaches, with Astroconformer reporting the best performance. The dataset establishes a common benchmark for future comparisons and advancements in rocket body classification tasks.


Bryndza at ClimateActivism 2024: Stance, Target and Hate Event Detection via Retrieval-Augmented GPT-4 and LLaMA

arXiv.org Artificial Intelligence

This study details our approach for the CASE 2024 Shared Task on Climate Activism Stance and Hate Event Detection, focusing on Hate Speech Detection, Hate Speech Target Identification, and Stance Detection as classification challenges. We explored the capability of Large Language Models (LLMs), particularly GPT-4, in zero- or few-shot settings enhanced by retrieval augmentation and re-ranking for Tweet classification. Our goal was to determine if LLMs could match or surpass traditional methods in this context. We conducted an ablation study with LLaMA for comparison, and our results indicate that our models significantly outperformed the baselines, securing second place in the Target Detection task. The code for our submission is available at https://github.com/NaiveNeuron/bryndza-case-2024


Universal NER: A Gold-Standard Multilingual Named Entity Recognition Benchmark

arXiv.org Artificial Intelligence

We introduce Universal NER (UNER), an open, community-driven project to develop gold-standard NER benchmarks in many languages. The overarching goal of UNER is to provide high-quality, cross-lingually consistent annotations to facilitate and standardize multilingual NER research. UNER v1 contains 18 datasets annotated with named entities in a cross-lingual consistent schema across 12 diverse languages. In this paper, we detail the dataset creation and composition of UNER; we also provide initial modeling baselines on both in-language and cross-lingual learning settings. We release the data, code, and fitted models to the public.


WikiGoldSK: Annotated Dataset, Baselines and Few-Shot Learning Experiments for Slovak Named Entity Recognition

arXiv.org Artificial Intelligence

Named Entity Recognition (NER) is a lower-level In this paper we focus on Slovak, a language Natural Language Processing (NLP) task in which of the Indo-European family, spoken by 5 million the aim is to both identify and classify named entity native speakers, which is still missing a manually expressions in text into a pre-defined set of annotated NER dataset of substantial size. To fill semantic types, such as Location, Organization or this gap, we propose the following contributions: Person (Goyal et al., 2018). It is a key component of many downstream NLP tasks, ranging from information We introduce a novel, manually annotated extraction, machine translation, question NER dataset called WikiGoldSK built by annotating answering to entity linking and co-reference resolution, articles sampled from Slovak Wikipedia among others. Since its introduction at and labeled with four entity classes. MUC-6 (Grishman and Sundheim, 1996), the task We evaluate a selection of multilingual NER has been studied extensively, usually as a form of baseline models on the presented dataset to token classification. In recent years, the advent compare its quality with that of existing silverstandard of pre-trained language models (PLMs) combined Slovak NER datasets.