Retrieval Augmented Large Language Model System for Comprehensive Drug Contraindications
- URL: http://arxiv.org/abs/2508.06145v1
- Date: Fri, 08 Aug 2025 09:09:03 GMT
- Title: Retrieval Augmented Large Language Model System for Comprehensive Drug Contraindications
- Authors: Byeonghun Bang, Jongsuk Yoon, Dong-Jin Chang, Seho Park, Yong Oh Lee,
- Abstract summary: The versatility of large language models (LLMs) has been explored across various sectors, but their application in healthcare poses challenges.<n>This study enhances the capability of LLMs to address contraindications effectively by implementing a Retrieval Augmented Generation (RAG) pipeline.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The versatility of large language models (LLMs) has been explored across various sectors, but their application in healthcare poses challenges, particularly in the domain of pharmaceutical contraindications where accurate and reliable information is required. This study enhances the capability of LLMs to address contraindications effectively by implementing a Retrieval Augmented Generation (RAG) pipeline. Utilizing OpenAI's GPT-4o-mini as the base model, and the text-embedding-3-small model for embeddings, our approach integrates Langchain to orchestrate a hybrid retrieval system with re-ranking. This system leverages Drug Utilization Review (DUR) data from public databases, focusing on contraindications for specific age groups, pregnancy, and concomitant drug use. The dataset includes 300 question-answer pairs across three categories, with baseline model accuracy ranging from 0.49 to 0.57. Post-integration of the RAG pipeline, we observed a significant improvement in model accuracy, achieving rates of 0.94, 0.87, and 0.89 for contraindications related to age groups, pregnancy, and concomitant drug use, respectively. The results indicate that augmenting LLMs with a RAG framework can substantially reduce uncertainty in prescription and drug intake decisions by providing more precise and reliable drug contraindication information.
Related papers
- A Federated and Parameter-Efficient Framework for Large Language Model Training in Medicine [59.78991974851707]
Large language models (LLMs) have demonstrated strong performance on medical benchmarks, including question answering and diagnosis.<n>Most medical LLMs are trained on data from a single institution, which faces limitations in generalizability and safety in heterogeneous systems.<n>We introduce the model-agnostic and parameter-efficient federated learning framework for adapting LLMs to medical applications.
arXiv Detail & Related papers (2026-01-29T18:48:21Z) - MixRx: Predicting Drug Combination Interactions with LLMs [0.0]
We evaluate the performance of 4 models, GPT-2, Mistral Instruct 2.0, and the fine-tuned counterparts.<n>Our results showed a potential for such an application, with the Mistral Instruct 2.0 Fine-Tuned model providing an average accuracy score on standard and perturbed datasets of 81.5%.
arXiv Detail & Related papers (2025-12-28T05:37:56Z) - DrugRAG: Enhancing Pharmacy LLM Performance Through A Novel Retrieval-Augmented Generation Pipeline [4.750574899254107]
We benchmarked eleven existing large language models (LLMs) with varying parameter sizes using a 141-question pharmacy dataset.<n>We developed a three-step retrieval-augmented generation (RAG) pipeline, DrugRAG, that retrieves structured drug knowledge from validated sources and augments model prompts with evidence-based context.<n>DrugRAG improved accuracy across all tested models, with gains ranging from 7 to 21 percentage points (e.g., Gemma 3 27B: 61% to 71%, Llama 3.1 8B: 46% to 67%) on the 141-item benchmark.
arXiv Detail & Related papers (2025-12-16T20:19:23Z) - CDrugRed: A Chinese Drug Recommendation Dataset for Discharge Medications in Metabolic Diseases [49.09102662968899]
We present CDrugRed, a first publicly available Chinese drug recommendation dataset focused on discharge medications for metabolic diseases.<n>The dataset includes 5,894 de-identified records from 3,190 patients, containing comprehensive information such as patient demographics, medical history, clinical course, and discharge diagnoses.<n>We assess the utility of CDrugRed by benchmarking several state-of-the-art large language models (LLMs) on the discharge medication recommendation task.
arXiv Detail & Related papers (2025-10-24T01:47:23Z) - A Hybrid Computational Intelligence Framework with Metaheuristic Optimization for Drug-Drug Interaction Prediction [0.8602553195689512]
Drug-drug interactions (DDIs) are a leading cause of preventable adverse events, often complicating treatment and increasing healthcare costs.<n>We propose an interpretable and efficient framework that blends modern machine learning with domain knowledge to improve DDI prediction.<n>Our approach combines two complementary embeddings - Mol2Vec, which captures fragment-level structural patterns, and SMILES-BERT, which learns contextual chemical features.
arXiv Detail & Related papers (2025-10-08T09:55:18Z) - RAG-based Architectures for Drug Side Effect Retrieval in LLMs [0.0]
Large Language Models (LLMs) offer promising conversational interfaces, but their inherent limitations hinder their reliability in specialized fields like pharmacovigilance.<n>We propose two architectures: Retrieval-Augmented Generation (RAG) and GraphRAG, which integrate comprehensive drug side effect knowledge into a Llama 3 8B language model.
arXiv Detail & Related papers (2025-07-18T11:20:52Z) - LLMs for Drug-Drug Interaction Prediction: A Comprehensive Comparison [3.2627279988912194]
Large Language Models (LLMs) have revolutionized various domains, but their potential in pharmaceutical research remains largely unexplored.<n>This study thoroughly investigates LLMs' capabilities in predicting drug-drug interactions (DDIs)<n>We evaluated 18 different LLMs, including proprietary models (GPT-4, Claude, Gemini) and open-source variants (from 1.5B to 72B parameters)<n>Fine-tuned LLMs demonstrated superior performance, with Phi-3.5 2.7B achieving a sensitivity of 0.978 in DDI prediction, with an accuracy of 0.919 on balanced datasets.
arXiv Detail & Related papers (2025-02-09T09:58:12Z) - Rx Strategist: Prescription Verification using LLM Agents System [0.0]
Rx Strategist uses knowledge graphs and different search strategies to enhance the power of Large Language Models (LLMs) inside an agentic framework.
This multifaceted technique allows for a multi-stage LLM pipeline and reliable information retrieval from a custom-built active ingredient database.
Our findings demonstrate that Rx Strategist surpasses many current LLMs, achieving performance comparable to that of a highly experienced clinical pharmacist.
arXiv Detail & Related papers (2024-09-05T11:42:26Z) - Large Language Model Distilling Medication Recommendation Model [58.94186280631342]
We harness the powerful semantic comprehension and input-agnostic characteristics of Large Language Models (LLMs)<n>Our research aims to transform existing medication recommendation methodologies using LLMs.<n>To mitigate this, we have developed a feature-level knowledge distillation technique, which transfers the LLM's proficiency to a more compact model.
arXiv Detail & Related papers (2024-02-05T08:25:22Z) - StratMed: Relevance Stratification between Biomedical Entities for
Sparsity on Medication Recommendation [9.296433860766165]
StratMed is a stratification strategy that overcomes the long-tailed problem and achieves fuller learning of sparse data.
It also utilizes a dual-property network to address the issue of mutual constraints on the safety and accuracy of medication combinations.
Our model reduces safety risk by 15.08%, improves accuracy by 0.36%, and reduces training time consumption by 81.66%.
arXiv Detail & Related papers (2023-08-31T14:59:32Z) - Drug Synergistic Combinations Predictions via Large-Scale Pre-Training
and Graph Structure Learning [82.93806087715507]
Drug combination therapy is a well-established strategy for disease treatment with better effectiveness and less safety degradation.
Deep learning models have emerged as an efficient way to discover synergistic combinations.
Our framework achieves state-of-the-art results in comparison with other deep learning-based methods.
arXiv Detail & Related papers (2023-01-14T15:07:43Z) - Neural Bandits for Data Mining: Searching for Dangerous Polypharmacy [63.135687276599114]
Some polypharmacies, deemed inappropriate, may be associated with adverse health outcomes such as death or hospitalization.
We propose the OptimNeuralTS strategy to efficiently mine claims datasets and build a predictive model of the association between drug combinations and health outcomes.
Our method can detect up to 72% of PIPs while maintaining an average precision score of 99% using 30 000 time steps.
arXiv Detail & Related papers (2022-12-10T03:43:23Z) - SSM-DTA: Breaking the Barriers of Data Scarcity in Drug-Target Affinity
Prediction [127.43571146741984]
Drug-Target Affinity (DTA) is of vital importance in early-stage drug discovery.
wet experiments remain the most reliable method, but they are time-consuming and resource-intensive.
Existing methods have primarily focused on developing techniques based on the available DTA data, without adequately addressing the data scarcity issue.
We present the SSM-DTA framework, which incorporates three simple yet highly effective strategies.
arXiv Detail & Related papers (2022-06-20T14:53:25Z) - Reinforcement learning and Bayesian data assimilation for model-informed
precision dosing in oncology [0.0]
Current strategies comprise model-informed dosing tables or are based on maximum a-posteriori estimates.
We propose three novel approaches for MIPD employing Bayesian data assimilation and/or reinforcement learning to control neutropenia.
These approaches have the potential to substantially reduce the incidence of life-threatening grade 4 and subtherapeutic grade 0 neutropenia.
arXiv Detail & Related papers (2020-06-01T16:38:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.