Goto

Collaborating Authors

 sequential training






Elastic Weight Consolidation for Knowledge Graph Continual Learning: An Empirical Evaluation

Jhajj, Gaganpreet, Lin, Fuhua

arXiv.org Artificial Intelligence

Knowledge graphs (KGs) require continual updates as new information emerges, but neural embedding models suffer from catastrophic forgetting when learning new tasks sequentially. We evaluate Elastic Weight Consolidation (EWC), a regularization-based continual learning method, on KG link prediction using TransE embeddings on FB15k-237. Across multiple experiments with five random seeds, we find that EWC reduces catastrophic forgetting from 12.62% to 6.85%, a 45.7% reduction compared to naive sequential training. We observe that the task partitioning strategy affects the magnitude of forgetting: relation-based partitioning (grouping triples by relation type) exhibits 9.8 percentage points higher forgetting than randomly partitioned tasks (12.62% vs 2.81%), suggesting that task construction influences evaluation outcomes. While focused on a single embedding model and dataset, our results demonstrate that EWC effectively mitigates catastrophic forgetting in KG continual learning and highlight the importance of evaluation protocol design.






KDH-MLTC: Knowledge Distillation for Healthcare Multi-Label Text Classification

Sakai, Hajar, Lam, Sarah S.

arXiv.org Artificial Intelligence

The increasing volume of healthcare textual data requires computationally efficient, yet highly accurate classification approaches able to handle the nuanced and complex nature of medical terminology. This research presents Knowledge Distillation for Healthcare Multi - Label Text Classification (KDH - MLTC), a framework leveraging model compr ession and Large Language Models (LLMs). The proposed approach addresses conventional healthcare Multi - Label Text Classification (MLTC) challenges by integrating knowledge distillation and sequential fine - tuning, subsequently optimized through Particle Swa rm Optimization (PSO) for hyperparameter tuning. KDH - MLTC transfers knowledge from a more complex teacher LLM ( i.e., BERT) to a lighter student LLM ( i.e., DistilBERT) through sequential training adapted to MLTC that preserves the teacher's learned information while significantly reducing computational requirements. As a result, the classification is enabled to be conducted locally, making it suitable for healthcare textual data characterized by sensitivity and, therefore, ensuring HIPAA compliance. The e xpe riments conducted on three medical literature datasets of different sizes, sampled from the Hallmark of Cancer (HoC) dataset, demonstrate that KDH - MLTC achieves superior performance compared to existing approaches, particularly for the largest dataset, reaching an F1 score of 82.70% 0.89%. Additionally, statistical validation and an ablation study ar e carried out, proving the robustness of KDH - MLTC. Furthermore, the PSO - based hyperparameter optimization process allow ed the identification of optimal configurations. The proposed approach contributes to healthcare text classification research, balancing efficiency requirements in resource - constrained healthcare settings with satisfactory accuracy demands.