Almanac: Retrieval-Augmented Language Models for Clinical Medicine
Zakka, Cyril, Chaurasia, Akash, Shad, Rohan, Dalal, Alex R., Kim, Jennifer L., Moor, Michael, Alexander, Kevin, Ashley, Euan, Boyd, Jack, Boyd, Kathleen, Hirsch, Karen, Langlotz, Curt, Nelson, Joanna, Hiesinger, William
–arXiv.org Artificial Intelligence
In recent years, language model pre-training has emerged as a powerful training paradigm in natural language processing (NLP) [1-4]. For a large number of these language models, performance improvements have been empirically observed to scale with model and dataset size, with the well-documented emergence of zero-shot capabilities and sample efficiency on a range of downstream NLP tasks [5-7]. However, due the nature of their training objective-- predicting the next token in a sentence--large language models (LLMs) can be prone to generating factually incorrect statements, a phenomenon commonly known as hallucination [8, 9]. More contentiously, many works have also demonstrated these models' ability to reproduce social biases, as well as generating statements reinforcing gender, racial, and religious stereotypes [10, 11]. In an effort to reduce these unwanted behaviors, several works have explored different ways of steering LLM outputs to more closely align with user-intent, including fine-tuning with human feedback [12, 13] and natural language prompt engineering [14, 15].
arXiv.org Artificial Intelligence
May-31-2023
- Country:
- Asia (0.28)
- North America > United States (0.46)
- Genre:
- Overview (0.93)
- Research Report
- Experimental Study (0.68)
- New Finding (1.00)
- Industry:
- Health & Medicine
- Diagnostic Medicine (0.93)
- Pharmaceuticals & Biotechnology (1.00)
- Therapeutic Area
- Cardiology/Vascular Diseases (1.00)
- Hematology (1.00)
- Immunology (1.00)
- Infections and Infectious Diseases (1.00)
- Neurology (1.00)
- Health & Medicine
- Technology: