Large Language Models are Few-Shot Health Learners
- URL: http://arxiv.org/abs/2305.15525v1
- Date: Wed, 24 May 2023 19:25:16 GMT
- Title: Large Language Models are Few-Shot Health Learners
- Authors: Xin Liu, Daniel McDuff, Geza Kovacs, Isaac Galatzer-Levy, Jacob
Sunshine, Jiening Zhan, Ming-Zher Poh, Shun Liao, Paolo Di Achille, Shwetak
Patel
- Abstract summary: Large language models (LLMs) can capture rich representations of concepts that are useful for real-world tasks.
Health applications require that models be grounded in numerical data.
We demonstrate that with only few-shot tuning, a large language model is capable of grounding various physiological and behavioral time-series data.
- Score: 8.549461354297677
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Large language models (LLMs) can capture rich representations of concepts
that are useful for real-world tasks. However, language alone is limited. While
existing LLMs excel at text-based inferences, health applications require that
models be grounded in numerical data (e.g., vital signs, laboratory values in
clinical domains; steps, movement in the wellness domain) that is not easily or
readily expressed as text in existing training corpus. We demonstrate that with
only few-shot tuning, a large language model is capable of grounding various
physiological and behavioral time-series data and making meaningful inferences
on numerous health tasks for both clinical and wellness contexts. Using data
from wearable and medical sensor recordings, we evaluate these capabilities on
the tasks of cardiac signal analysis, physical activity recognition, metabolic
calculation (e.g., calories burned), and estimation of stress reports and
mental health screeners.
Related papers
- Scaling Wearable Foundation Models [54.93979158708164]
We investigate the scaling properties of sensor foundation models across compute, data, and model size.
Using a dataset of up to 40 million hours of in-situ heart rate, heart rate variability, electrodermal activity, accelerometer, skin temperature, and altimeter per-minute data from over 165,000 people, we create LSM.
Our results establish the scaling laws of LSM for tasks such as imputation, extrapolation, both across time and sensor modalities.
arXiv Detail & Related papers (2024-10-17T15:08:21Z) - Transfer Learning with Clinical Concept Embeddings from Large Language Models [4.838020198075334]
Large Language Models (LLMs) show significant potential of capturing the semantic meaning of clinical concepts.
This study analyzed electronic health records from two large healthcare systems to assess the impact of semantic embeddings on local, shared, and transfer learning models.
arXiv Detail & Related papers (2024-09-20T20:50:55Z) - Medical Vision-Language Pre-Training for Brain Abnormalities [96.1408455065347]
We show how to automatically collect medical image-text aligned data for pretraining from public resources such as PubMed.
In particular, we present a pipeline that streamlines the pre-training process by initially collecting a large brain image-text dataset.
We also investigate the unique challenge of mapping subfigures to subcaptions in the medical domain.
arXiv Detail & Related papers (2024-04-27T05:03:42Z) - Redefining Digital Health Interfaces with Large Language Models [69.02059202720073]
Large Language Models (LLMs) have emerged as general-purpose models with the ability to process complex information.
We show how LLMs can provide a novel interface between clinicians and digital technologies.
We develop a new prognostic tool using automated machine learning.
arXiv Detail & Related papers (2023-10-05T14:18:40Z) - Interpretable Medical Diagnostics with Structured Data Extraction by
Large Language Models [59.89454513692417]
Tabular data is often hidden in text, particularly in medical diagnostic reports.
We propose a novel, simple, and effective methodology for extracting structured tabular data from textual medical reports, called TEMED-LLM.
We demonstrate that our approach significantly outperforms state-of-the-art text classification models in medical diagnostics.
arXiv Detail & Related papers (2023-06-08T09:12:28Z) - Do We Still Need Clinical Language Models? [15.023633270864675]
We show that relatively small specialized clinical models substantially outperform all in-context learning approaches.
We release the code and the models used under the PhysioNet Credentialed Health Data license and data use agreement.
arXiv Detail & Related papers (2023-02-16T05:08:34Z) - A Multi-View Joint Learning Framework for Embedding Clinical Codes and
Text Using Graph Neural Networks [23.06795121693656]
We propose a framework that learns from codes and text to combine the availability and forward-looking nature of text and better performance of ICD codes.
Our approach uses a Graph Neural Network (GNN) to process ICD codes, and Bi-LSTM to process text.
In experiments using planned surgical procedure text, our model outperforms BERT models fine-tuned to clinical data.
arXiv Detail & Related papers (2023-01-27T09:19:03Z) - CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark [51.38557174322772]
We present the first Chinese Biomedical Language Understanding Evaluation benchmark.
It is a collection of natural language understanding tasks including named entity recognition, information extraction, clinical diagnosis normalization, single-sentence/sentence-pair classification.
We report empirical results with the current 11 pre-trained Chinese models, and experimental results show that state-of-the-art neural models perform by far worse than the human ceiling.
arXiv Detail & Related papers (2021-06-15T12:25:30Z) - Automated Coding of Under-Studied Medical Concept Domains: Linking
Physical Activity Reports to the International Classification of Functioning,
Disability, and Health [22.196642357767338]
Many domains of medical concepts lack well-developed terminologies that can support effective coding of medical text.
We present a framework for developing natural language processing (NLP) technologies for automated coding of under-studied types of medical information.
arXiv Detail & Related papers (2020-11-27T20:02:59Z) - Self-supervised transfer learning of physiological representations from
free-living wearable data [12.863826659440026]
We present a novel self-supervised representation learning method using activity and heart rate (HR) signals without semantic labels.
We evaluate our model in the largest free-living combined-sensing dataset (comprising >280k hours of wrist accelerometer & wearable ECG data)
arXiv Detail & Related papers (2020-11-18T23:21:34Z) - Domain-Specific Language Model Pretraining for Biomedical Natural
Language Processing [73.37262264915739]
We show that for domains with abundant unlabeled text, such as biomedicine, pretraining language models from scratch results in substantial gains.
Our experiments show that domain-specific pretraining serves as a solid foundation for a wide range of biomedical NLP tasks.
arXiv Detail & Related papers (2020-07-31T00:04:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.