Supervised Fine-Tuning or In-Context Learning? Evaluating LLMs for Clinical NER
–arXiv.org Artificial Intelligence
We study clinical Named Entity Recognition (NER) on the CADEC corpus and compare three families of approaches: (i) BERT-style encoders (BERT Base, BioClinicalBERT, RoBERTa-large), (ii) GPT-4o used with few-shot in-context learning (ICL) under simple vs.\ complex prompts, and (iii) GPT-4o with supervised fine-tuning (SFT). All models are evaluated on standard NER metrics over CADEC's five entity types (ADR, Drug, Disease, Symptom, Finding). RoBERTa-large and BioClinicalBERT offer limited improvements over BERT Base, showing the limit of these family of models. Among LLM settings, simple ICL outperforms a longer, instruction-heavy prompt, and SFT achieves the strongest overall performance (F1 $\approx$ 87.1%), albeit with higher cost. We find that the LLM achieve higher accuracy on simplified tasks, restricting classification to two labels.
arXiv.org Artificial Intelligence
Oct-28-2025
- Country:
- Europe > Netherlands
- South Holland > Leiden (0.04)
- North America > Canada (0.04)
- Europe > Netherlands
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine
- Consumer Health (0.68)
- Health Care Technology > Medical Record (0.47)
- Pharmaceuticals & Biotechnology (0.68)
- Therapeutic Area > Neurology (0.46)
- Information Technology > Security & Privacy (0.68)
- Health & Medicine
- Technology: