Metadata Might Make Language Models Better

Beelen, Kaspar, van Strien, Daniel

arXiv.org Artificial Intelligence 

Using 19th-century newspapers as a case study, we extend the time-masking approach proposed by Rosin et al. [2022] and compare different strategies for inserting temporal, political and geographical information into a Masked Language Model. After fine-tuning several DistilBERT on enhanced input data, we provide a systematic evaluation of these models on a set of evaluation tasks: pseudo-perplexity, metadata mask-filling and supervised classification. We find that showing relevant metadata to a language model has a beneficial impact and may even produce more robust and fairer models.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found