Meta-Pretraining for Zero-Shot Cross-Lingual Named Entity Recognition in Low-Resource Philippine Languages
Africa, David Demitri, Salhan, Suchir, Weiss, Yuval, Buttery, Paula, Martinez, Richard Diehl
–arXiv.org Artificial Intelligence
Named-entity recognition (NER) in low-resource languages is usually tackled by finetuning very large multilingual LMs, an option that is often infeasible in memory- or latency-constrained settings. We ask whether small decoder LMs can be pretrained so that they adapt quickly and transfer zero-shot to languages unseen during pretraining. To this end we replace part of the autoregressive objective with first-order model-agnostic meta-learning (MAML). Tagalog and Cebuano are typologically similar yet structurally different in their actor/non-actor voice systems, and hence serve as a challenging test-bed. Across four model sizes (11 M - 570 M) MAML lifts zero-shot micro-F1 by 2-6 pp under head-only tuning and 1-3 pp after full tuning, while cutting convergence time by up to 8%. Gains are largest for single-token person entities that co-occur with Tagalog case particles si/ni, highlighting the importance of surface anchors.
arXiv.org Artificial Intelligence
Oct-7-2025
- Country:
- Africa (0.41)
- Asia
- China > Hong Kong (0.04)
- Indonesia > Bali (0.04)
- Philippines > Luzon
- National Capital Region > City of Manila (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Europe
- Austria > Vienna (0.14)
- Germany > Berlin (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- California > San Francisco County
- San Francisco (0.14)
- Florida > Miami-Dade County
- Miami (0.04)
- California > San Francisco County
- Mexico > Mexico City
- Genre:
- Research Report (0.82)
- Technology: