fine-tuning technique
- North America > United States > Texas (0.04)
- South America > Colombia > Meta Department > Villavicencio (0.04)
- Oceania > Australia > New South Wales (0.04)
- (11 more...)
- North America > United States > Texas (0.04)
- South America > Colombia > Meta Department > Villavicencio (0.04)
- Oceania > Australia > New South Wales (0.04)
- (11 more...)
ZhiFangDanTai: Fine-tuning Graph-based Retrieval-Augmented Generation Model for Traditional Chinese Medicine Formula
Zhang, ZiXuan, Hao, Bowen, Li, Yingjie, Yin, Hongzhi
Traditional Chinese Medicine (TCM) formulas play a significant role in treating epidemics and complex diseases. Existing models for TCM utilize traditional algorithms or deep learning techniques to analyze formula relationships, yet lack comprehensive results, such as complete formula compositions and detailed explanations. Although recent efforts have used TCM instruction datasets to fine-tune Large Language Models (LLMs) for explainable formula generation, existing datasets lack sufficient details, such as the roles of the formula's sovereign, minister, assistant, courier; efficacy; contraindications; tongue and pulse diagnosis-limiting the depth of model outputs. To address these challenges, we propose ZhiFangDanTai, a framework combining Graph-based Retrieval-Augmented Generation (GraphRAG) with LLM fine-tuning. ZhiFangDanTai uses GraphRAG to retrieve and synthesize structured TCM knowledge into concise summaries, while also constructing an enhanced instruction dataset to improve LLMs' ability to integrate retrieved information. Furthermore, we provide novel theoretical proofs demonstrating that integrating GraphRAG with fine-tuning techniques can reduce generalization error and hallucination rates in the TCM formula task. Experimental results on both collected and clinical datasets demonstrate that ZhiFangDanTai achieves significant improvements over state-of-the-art models. Our model is open-sourced at https://huggingface.co/tczzx6/ZhiFangDanTai1.0.
- Europe > Netherlands > South Holland > Leiden (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Oceania > Australia > Queensland > Brisbane (0.04)
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Education (1.00)
- Health & Medicine > Therapeutic Area > Cardiology/Vascular Diseases (0.93)
- Health & Medicine > Consumer Health (0.92)
- (3 more...)
Reinforcement learning fine-tuning of language model for instruction following and math reasoning
This study investigates the effectiveness of reinforcement learning (RL) fine-tuning techniques on a compact language model (Qwen2.5-0.5B Base) for two challenging tasks: instruction following and mathematical reasoning. We compare supervised fine-tuning (SFT), Direct Preference Optimization (DPO) using preference-labeled data, and Reinforce Leave-One-Out (RLOO) with reward models. Our experiments show that RLOO with DeBERTa reward modeling achieves the best alignment, while DPO provides strong and consistent results. For math reasoing tasks, synthetic data augmentation and best-of-N sampling with an external verifier significantly improve accuracy, showing the potential of combining fine-tuning with inference-time tools. This study highlights key trade-offs and practical strategies for training lightweight, task-aligned small-scale language models.
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.86)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.70)
EMRModel: A Large Language Model for Extracting Medical Consultation Dialogues into Structured Medical Records
Zhao, Shuguang, Feng, Qiangzhong, He, Zhiyang, Sun, Peipei, Wang, Yingying, Tao, Xiaodong, Lu, Xiaoliang, Cheng, Mei, Wu, Xinyue, Wang, Yanyan, Liang, Wei
Medical consultation dialogues contain critical clinical information, yet their unstructured nature hinders effective utilization in diagnosis and treatment. Traditional methods, relying on rule-based or shallow machine learning techniques, struggle to capture deep and implicit semantics. Recently, large pre-trained language models and Low-Rank Adaptation (LoRA), a lightweight fine-tuning method, have shown promise for structured information extraction. We propose EMRModel, a novel approach that integrates LoRA-based fine-tuning with code-style prompt design, aiming to efficiently convert medical consultation dialogues into structured electronic medical records (EMRs). Additionally, we construct a high-quality, realistically grounded dataset of medical consultation dialogues with detailed annotations. Furthermore, we introduce a fine-grained evaluation benchmark for medical consultation information extraction and provide a systematic evaluation methodology, advancing the optimization of medical natural language processing (NLP) models. Experimental results show EMRModel achieves an F1 score of 88.1%, improving by49.5% over standard pre-trained models. Compared to traditional LoRA fine-tuning methods, our model shows superior performance, highlighting its effectiveness in structured medical record extraction tasks.
- North America > United States > California > Los Angeles County > Los Angeles (0.28)
- Asia > China > Anhui Province > Hefei (0.04)
- North America > United States > Montana > Roosevelt County (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Information Extraction (0.75)
Pre-train and Fine-tune: Recommenders as Large Models
Jiang, Zhenhao, Chen, Chenghao, Feng, Hao, Yang, Yu, Liu, Jin, Zhang, Jie, Jia, Jia, Hu, Ning
In reality, users have different interests in different periods, regions, scenes, etc. Such changes in interest are so drastic that they are difficult to be captured by recommenders. Existing multi-domain learning can alleviate this problem. However, the structure of the industrial recommendation system is complex, the amount of data is huge, and the training cost is extremely high, so it is difficult to modify the structure of the industrial recommender and re-train it. To fill this gap, we consider recommenders as large pre-trained models and fine-tune them. We first propose the theory of the information bottleneck for fine-tuning and present an explanation for the fine-tuning technique in recommenders. To tailor for recommendation, we design an information-aware adaptive kernel (IAK) technique to fine-tune the pre-trained recommender. Specifically, we define fine-tuning as two phases: knowledge compression and knowledge matching and let the training stage of IAK explicitly approximate these two phases. Our proposed approach designed from the essence of fine-tuning is well interpretable. Extensive online and offline experiments show the superiority of our proposed method. Besides, we also share unique and important lessons we learned when deploying the method in a large-scale online platform. We also present the potential issues of fine-tuning techniques in recommendation systems and the corresponding solutions. The recommender with IAK technique has been deployed on the homepage of a billion-scale online food platform for several months and has yielded considerable profits in our business.
- Oceania > Australia > New South Wales > Sydney (0.05)
- Asia > China > Zhejiang Province > Hangzhou (0.05)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- (2 more...)
PyraNet: A Large Scale Hierarchical Verilog Dataset
Nadimi, Bardia, Boutaib, Ghali Omar, Zheng, Hao
Recently, there has been a growing interest in leveraging Large Language Models for Verilog code generation. However, the current quality of the generated Verilog code remains suboptimal. This is largely due to the absence of well-defined, well-organized datasets with high-quality samples, as well as a lack of innovative fine-tuning methods and models specifically trained on Verilog. In this paper, we introduce a novel open-source dataset and a corresponding fine-tuning technique, which utilizes a multi-layered structure that we refer to as PyraNet. Our experiments demonstrate that employing the proposed dataset and fine-tuning approach leads to a more accurate fine-tuned model, producing syntactically and functionally correct Verilog code. The evaluation results show improvements by up-to $32.6\%$ in comparison to the CodeLlama-7B baseline model and up-to $16.7\%$ in comparison to the state-of-the-art models using VerilogEval evaluation platform.
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- North America > United States > Florida > Hillsborough County > University (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (3 more...)
- Research Report > New Finding (0.48)
- Research Report > Promising Solution (0.34)
Fine-Tuning LLMs for Reliable Medical Question-Answering Services
Anaissi, Ali, Braytee, Ali, Akram, Junaid
We present an advanced approach to medical question-answering (QA) services, using fine-tuned Large Language Models (LLMs) to improve the accuracy and reliability of healthcare information. Our study focuses on optimizing models like LLaMA-2 and Mistral, which have shown great promise in delivering precise, reliable medical answers. By leveraging comprehensive datasets, we applied fine-tuning techniques such as rsDoRA+ and ReRAG. rsDoRA+ enhances model performance through a combination of decomposed model weights, varied learning rates for low-rank matrices, and rank stabilization, leading to improved efficiency. ReRAG, which integrates retrieval on demand and question rewriting, further refines the accuracy of the responses. This approach enables healthcare providers to access fast, dependable information, aiding in more efficient decision-making and fostering greater patient trust. Our work highlights the potential of fine-tuned LLMs to significantly improve the quality and accessibility of medical information services, ultimately contributing to better healthcare outcomes for all.
- Health & Medicine > Therapeutic Area (0.47)
- Health & Medicine > Consumer Health (0.46)
Efficient Fine-Tuning of Large Language Models for Automated Medical Documentation
Leong, Hui Yi, Gao, Yi Fan, Shuai, Ji, Pamuksuz, Uktu
Scientific research indicates that for every hour spent in direct patient care, physicians spend nearly two additional hours on administrative tasks, particularly on electronic health records (EHRs) and desk work. This excessive administrative burden not only reduces the time available for patient care but also contributes to physician burnout and inefficiencies in healthcare delivery. To address these challenges, this study introduces MediGen, a fine-tuned large language model (LLM) designed to automate the generation of medical reports from medical dialogues. By leveraging state-of-the-art methodologies for fine-tuning open-source pretrained models, including LLaMA3-8B, MediGen achieves high accuracy in transcribing and summarizing clinical interactions. The fine-tuned LLaMA3-8B model demonstrated promising results, achieving a ROUGE score of 58% and a BERTScore-F1 of 72%, indicating its effectiveness in generating accurate and clinically relevant medical reports. These findings suggest that MediGen has the potential to significantly reduce the administrative workload on physicians, improving both healthcare efficiency and physician well-being.
HyperLoader: Integrating Hypernetwork-Based LoRA and Adapter Layers into Multi-Task Transformers for Sequence Labelling
Ortiz-Barajas, Jesus-German, Gomez-Adorno, Helena, Solorio, Thamar
We use the encoder-decoder T5 model only a small number of parameters is updated to (Raffel et al., 2020) for all experiments to take a downstream task (Houlsby et al., 2019; Stickland advantage of modelling the tasks as sequence-tosequence and Murray, 2019; Karimi Mahabadi et al., tasks. We test our model in seven datasets 2021a). These methods aim to achieve comparable from two Sequence Labelling tasks. The first task performance to full fine-tuning by updating as few is Named Entity Recognition, a valuable tool in parameters as possible. However, a less studied research various real-world scenarios in the era of large language direction related to these methods is whether models such as healthcare and medical research one can perform better than full fine-tuning with (Raza et al., 2022; Hu et al., 2024), Finance fewer parameters (Mao et al., 2022).
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Singapore (0.04)
- North America > Dominican Republic (0.04)
- (8 more...)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.55)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.46)
- Information Technology > Artificial Intelligence > Natural Language > Information Retrieval (0.35)