Neural Medication Extraction: A Comparison of Recent Models in
Supervised and Semi-supervised Learning Settings
- URL: http://arxiv.org/abs/2110.10213v1
- Date: Tue, 19 Oct 2021 19:23:38 GMT
- Title: Neural Medication Extraction: A Comparison of Recent Models in
Supervised and Semi-supervised Learning Settings
- Authors: Ali Can Kocabiyikoglu, Fran\c{c}ois Portet, Raheel Qader, Jean-Marc
Babouchkine
- Abstract summary: Drug prescriptions are essential information that must be encoded in electronic medical records.
This is why the medication extraction task has emerged.
We present an independent and comprehensive evaluation of state-of-the-art neural architectures on the I2B2 medical prescription extraction task.
- Score: 0.751289645756884
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Drug prescriptions are essential information that must be encoded in
electronic medical records. However, much of this information is hidden within
free-text reports. This is why the medication extraction task has emerged. To
date, most of the research effort has focused on small amount of data and has
only recently considered deep learning methods. In this paper, we present an
independent and comprehensive evaluation of state-of-the-art neural
architectures on the I2B2 medical prescription extraction task both in the
supervised and semi-supervised settings. The study shows the very competitive
performance of simple DNN models on the task as well as the high interest of
pre-trained models. Adapting the latter models on the I2B2 dataset enables to
push medication extraction performances above the state-of-the-art. Finally,
the study also confirms that semi-supervised techniques are promising to
leverage large amounts of unlabeled data in particular in low resource setting
when labeled data is too costly to acquire.
Related papers
- Weakly supervised deep learning model with size constraint for prostate cancer detection in multiparametric MRI and generalization to unseen domains [0.90668179713299]
We show that the model achieves on-par performance with strong fully supervised baseline models.
We also observe a performance decrease for both fully supervised and weakly supervised models when tested on unseen data domains.
arXiv Detail & Related papers (2024-11-04T12:24:33Z) - SMILES-Mamba: Chemical Mamba Foundation Models for Drug ADMET Prediction [16.189335444981353]
Predicting the absorption, distribution, metabolism, excretion, and toxicity of small-molecule drugs is critical for ensuring safety and efficacy.
We propose a two-stage model that leverages both unlabeled and labeled data through a combination of self-supervised pretraining and fine-tuning strategies.
Our results demonstrate that SMILES-Mamba exhibits competitive performance across 22 ADMET datasets, achieving the highest score in 14 tasks.
arXiv Detail & Related papers (2024-08-11T04:53:12Z) - Extracting Training Data from Unconditional Diffusion Models [76.85077961718875]
diffusion probabilistic models (DPMs) are being employed as mainstream models for generative artificial intelligence (AI)
We aim to establish a theoretical understanding of memorization in DPMs with 1) a memorization metric for theoretical analysis, 2) an analysis of conditional memorization with informative and random labels, and 3) two better evaluation metrics for measuring memorization.
Based on the theoretical analysis, we propose a novel data extraction method called textbfSurrogate condItional Data Extraction (SIDE) that leverages a trained on generated data as a surrogate condition to extract training data directly from unconditional diffusion models.
arXiv Detail & Related papers (2024-06-18T16:20:12Z) - Drug Synergistic Combinations Predictions via Large-Scale Pre-Training
and Graph Structure Learning [82.93806087715507]
Drug combination therapy is a well-established strategy for disease treatment with better effectiveness and less safety degradation.
Deep learning models have emerged as an efficient way to discover synergistic combinations.
Our framework achieves state-of-the-art results in comparison with other deep learning-based methods.
arXiv Detail & Related papers (2023-01-14T15:07:43Z) - SSM-DTA: Breaking the Barriers of Data Scarcity in Drug-Target Affinity
Prediction [127.43571146741984]
Drug-Target Affinity (DTA) is of vital importance in early-stage drug discovery.
wet experiments remain the most reliable method, but they are time-consuming and resource-intensive.
Existing methods have primarily focused on developing techniques based on the available DTA data, without adequately addressing the data scarcity issue.
We present the SSM-DTA framework, which incorporates three simple yet highly effective strategies.
arXiv Detail & Related papers (2022-06-20T14:53:25Z) - DrugOOD: Out-of-Distribution (OOD) Dataset Curator and Benchmark for
AI-aided Drug Discovery -- A Focus on Affinity Prediction Problems with Noise
Annotations [90.27736364704108]
We present DrugOOD, a systematic OOD dataset curator and benchmark for AI-aided drug discovery.
DrugOOD comes with an open-source Python package that fully automates benchmarking processes.
We focus on one of the most crucial problems in AIDD: drug target binding affinity prediction.
arXiv Detail & Related papers (2022-01-24T12:32:48Z) - Selection of pseudo-annotated data for adverse drug reaction
classification across drug groups [12.259552039796027]
We assess the robustness of state-of-the-art neural architectures across different drug groups.
We investigate several strategies to use pseudo-labeled data in addition to a manually annotated train set.
arXiv Detail & Related papers (2021-11-24T13:11:05Z) - On the diminishing return of labeling clinical reports [2.1431637042179683]
We show that performant medical NLP models may be obtained with small amount of labeled data.
We show quantitatively the effect of training data size on a fixed test set composed of two of the largest public chest x-ray radiology report datasets.
arXiv Detail & Related papers (2020-10-27T19:51:04Z) - Dual-Teacher: Integrating Intra-domain and Inter-domain Teachers for
Annotation-efficient Cardiac Segmentation [65.81546955181781]
We propose a novel semi-supervised domain adaptation approach, namely Dual-Teacher.
The student model learns the knowledge of unlabeled target data and labeled source data by two teacher models.
We demonstrate that our approach is able to concurrently utilize unlabeled data and cross-modality data with superior performance.
arXiv Detail & Related papers (2020-07-13T10:00:44Z) - Confident Coreset for Active Learning in Medical Image Analysis [57.436224561482966]
We propose a novel active learning method, confident coreset, which considers both uncertainty and distribution for effectively selecting informative samples.
By comparative experiments on two medical image analysis tasks, we show that our method outperforms other active learning methods.
arXiv Detail & Related papers (2020-04-05T13:46:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.