Hippocrates: An Open-Source Framework for Advancing Large Language   Models in Healthcare
        - URL: http://arxiv.org/abs/2404.16621v1
- Date: Thu, 25 Apr 2024 14:06:37 GMT
- Title: Hippocrates: An Open-Source Framework for Advancing Large Language   Models in Healthcare
- Authors: Emre Can Acikgoz, Osman Batur İnce, Rayene Bench, Arda Anıl Boz, İlker Kesen, Aykut Erdem, Erkut Erdem, 
- Abstract summary: We present Hippocrates, an open-source framework specifically developed for the medical domain.
This open approach is designed to stimulate collaborative research, allowing the community to build upon, refine, and rigorously evaluate medical LLMs.
Also, we introduce Hippo, a family of 7B models tailored for the medical domain, fine-tuned from Mistral and LLaMA2 through continual pre-training, instruction tuning, and reinforcement learning from human and AI feedback.
- Score: 12.861529503960393
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract:   The integration of Large Language Models (LLMs) into healthcare promises to transform medical diagnostics, research, and patient care. Yet, the progression of medical LLMs faces obstacles such as complex training requirements, rigorous evaluation demands, and the dominance of proprietary models that restrict academic exploration. Transparent, comprehensive access to LLM resources is essential for advancing the field, fostering reproducibility, and encouraging innovation in healthcare AI. We present Hippocrates, an open-source LLM framework specifically developed for the medical domain. In stark contrast to previous efforts, it offers unrestricted access to its training datasets, codebase, checkpoints, and evaluation protocols. This open approach is designed to stimulate collaborative research, allowing the community to build upon, refine, and rigorously evaluate medical LLMs within a transparent ecosystem. Also, we introduce Hippo, a family of 7B models tailored for the medical domain, fine-tuned from Mistral and LLaMA2 through continual pre-training, instruction tuning, and reinforcement learning from human and AI feedback. Our models outperform existing open medical LLMs models by a large-margin, even surpassing models with 70B parameters. Through Hippocrates, we aspire to unlock the full potential of LLMs not just to advance medical knowledge and patient care but also to democratize the benefits of AI research in healthcare, making them available across the globe. 
 
      
        Related papers
        - MAM: Modular Multi-Agent Framework for Multi-Modal Medical Diagnosis via   Role-Specialized Collaboration [57.98393950821579]
 We introduce the Modular Multi-Agent Framework for Multi-Modal Medical Diagnosis (MAM)<n>Inspired by our empirical findings, MAM decomposes the medical diagnostic process into specialized roles: a General Practitioner, Specialist Team, Radiologist, Medical Assistant, and Director.<n>This modular and collaborative framework enables efficient knowledge updates and leverages existing medical LLMs and knowledge bases.
 arXiv  Detail & Related papers  (2025-06-24T17:52:43Z)
- Open Foundation Models in Healthcare: Challenges, Paradoxes, and   Opportunities with GenAI Driven Personalized Prescription [3.9083860193371938]
 In response to the success of proprietary Large Language Models (LLMs) such as OpenAI's GPT-4, there is a growing interest in developing open, non-proprietary AI foundation models (AIFMs)
Despite their inability to match the refined functionalities of their proprietary counterparts, open models hold immense potential to revolutionize healthcare applications.
 arXiv  Detail & Related papers  (2025-02-04T19:16:56Z)
- RuleAlign: Making Large Language Models Better Physicians with   Diagnostic Rule Alignment [54.91736546490813]
 We introduce the RuleAlign framework, designed to align Large Language Models with specific diagnostic rules.
We develop a medical dialogue dataset comprising rule-based communications between patients and physicians.
 Experimental results demonstrate the effectiveness of the proposed approach.
 arXiv  Detail & Related papers  (2024-08-22T17:44:40Z)
- A Survey on Large Language Models from General Purpose to Medical   Applications: Datasets, Methodologies, and Evaluations [5.265452667976959]
 This survey systematically summarizes how to train medical LLMs based on open-source general LLMs.
It covers (a) how to acquire training corpus and construct customized medical training sets, (b) how to choose an appropriate training paradigm, and (d) existing challenges and promising research directions.
 arXiv  Detail & Related papers  (2024-06-14T02:42:20Z)
- A Survey on Medical Large Language Models: Technology, Application,   Trustworthiness, and Future Directions [31.04135502285516]
 Large language models (LLMs) have received substantial attention due to their impressive capabilities for generating and understanding human-level language.
LLMs have emerged as an innovative and powerful adjunct in the medical field, transforming traditional practices and heralding a new era of enhanced healthcare services.
 arXiv  Detail & Related papers  (2024-06-06T03:15:13Z)
- Aloe: A Family of Fine-tuned Open Healthcare LLMs [0.0]
 We introduce the Aloe family, a set of open medical LLMs highly competitive within its scale range.
 Aloe models undergo an alignment phase, becoming one of the first few policy-aligned open healthcare LLMs.
To explore the limits of current LLMs in inference, we study several advanced prompt engineering strategies.
 arXiv  Detail & Related papers  (2024-05-03T07:14:07Z)
- AI Hospital: Benchmarking Large Language Models in a Multi-agent Medical   Interaction Simulator [69.51568871044454]
 We introduce textbfAI Hospital, a framework simulating dynamic medical interactions between emphDoctor as player and NPCs.
This setup allows for realistic assessments of LLMs in clinical scenarios.
We develop the Multi-View Medical Evaluation benchmark, utilizing high-quality Chinese medical records and NPCs.
 arXiv  Detail & Related papers  (2024-02-15T06:46:48Z)
- Large Language Model Distilling Medication Recommendation Model [61.89754499292561]
 We harness the powerful semantic comprehension and input-agnostic characteristics of Large Language Models (LLMs)
Our research aims to transform existing medication recommendation methodologies using LLMs.
To mitigate this, we have developed a feature-level knowledge distillation technique, which transfers the LLM's proficiency to a more compact model.
 arXiv  Detail & Related papers  (2024-02-05T08:25:22Z)
- MEDITRON-70B: Scaling Medical Pretraining for Large Language Models [91.25119823784705]
 Large language models (LLMs) can potentially democratize access to medical knowledge.
We release MEDITRON: a suite of open-source LLMs with 7B and 70B parameters adapted to the medical domain.
 arXiv  Detail & Related papers  (2023-11-27T18:49:43Z)
- ChiMed-GPT: A Chinese Medical Large Language Model with Full Training   Regime and Better Alignment to Human Preferences [51.66185471742271]
 We propose ChiMed-GPT, a benchmark LLM designed explicitly for Chinese medical domain.
ChiMed-GPT undergoes a comprehensive training regime with pre-training, SFT, and RLHF.
We analyze possible biases through prompting ChiMed-GPT to perform attitude scales regarding discrimination of patients.
 arXiv  Detail & Related papers  (2023-11-10T12:25:32Z)
- A Survey of Large Language Models in Medicine: Progress, Application,   and Challenge [85.09998659355038]
 Large language models (LLMs) have received substantial attention due to their capabilities for understanding and generating human language.
This review aims to provide a detailed overview of the development and deployment of LLMs in medicine.
 arXiv  Detail & Related papers  (2023-11-09T02:55:58Z)
- Large Language Models Illuminate a Progressive Pathway to Artificial
  Healthcare Assistant: A Review [16.008511195589925]
 Large language models (LLMs) have shown promising capabilities in mimicking human-level language comprehension and reasoning.
This paper provides a comprehensive review on the applications and implications of LLMs in medicine.
 arXiv  Detail & Related papers  (2023-11-03T13:51:36Z)
- MedAlpaca -- An Open-Source Collection of Medical Conversational AI
  Models and Training Data [40.97474177100237]
 Large language models (LLMs) hold considerable promise for improving medical, diagnostics, patient care, and education.
Yet, there is an urgent need for open-source models that can be deployed on-premises to safeguard patient privacy.
We present an innovative dataset consisting of over 160,000 entries, specifically crafted to fine-tune LLMs for effective medical applications.
 arXiv  Detail & Related papers  (2023-04-14T11:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.