Intertextual Parallel Detection in Biblical Hebrew: A Transformer-Based Benchmark
–arXiv.org Artificial Intelligence
Identifying parallel passages in biblical Hebrew (BH) is central to biblical scholarship for understanding intertextual relationships. Traditional methods rely on manual comparison, a labor-intensive process prone to human error. This study evaluates the potential of pre-trained transformer-based language models, including E5, AlephBERT, MPNet, and LaBSE, for detecting textual parallels in the Hebrew Bible. Focusing on known parallels between Samuel/Kings and Chronicles, I assessed each model's capability to generate word embeddings distinguishing parallel from non-parallel passages. Using cosine similarity and Wasserstein Distance measures, I found that E5 and AlephBERT show promise; E5 excels in parallel detection, while AlephBERT demonstrates stronger non-parallel differentiation. These findings indicate that pre-trained models can enhance the efficiency and accuracy of detecting intertextual parallels in ancient texts, suggesting broader applications for ancient language studies.
arXiv.org Artificial Intelligence
Jul-2-2025
- Country:
- Asia > Middle East
- Israel > Jerusalem District > Jerusalem (0.04)
- Europe > Netherlands
- North Holland > Amsterdam (0.04)
- North America > United States
- New York (0.04)
- Asia > Middle East
- Genre:
- Research Report (1.00)
- Technology: