The Efficiency of Pre-training with Objective Masking in Pseudo Labeling for Semi-Supervised Text Classification
Hatefi, Arezoo, Vu, Xuan-Son, Bhuyan, Monowar, Drewes, Frank
–arXiv.org Artificial Intelligence
We extend and study a semi-supervised model for text classification proposed earlier by Hatefi et al. for classification tasks in which document classes are described by a small number of gold-labeled examples, while the majority of training examples is unlabeled. The model leverages the teacher-student architecture of Meta Pseudo Labels in which a ''teacher'' generates labels for originally unlabeled training data to train the ''student'' and updates its own model iteratively based on the performance of the student on the gold-labeled portion of the data. We extend the original model of Hatefi et al. by an unsupervised pre-training phase based on objective masking, and conduct in-depth performance evaluations of the original model, our extension, and various independent baselines. Experiments are performed using three different datasets in two different languages (English and Swedish).
arXiv.org Artificial Intelligence
May-13-2025
- Country:
- Asia
- China (0.04)
- Middle East > Jordan (0.04)
- Europe
- Sweden (0.04)
- United Kingdom > England (0.04)
- North America
- Canada > Ontario
- Toronto (0.04)
- Dominican Republic (0.04)
- United States > Washington
- King County > Seattle (0.04)
- Canada > Ontario
- Asia
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Education (0.93)
- Health & Medicine (0.93)
- Media > News (0.67)
- Technology: