Neural translation and automated recognition of ICD10 medical entities
from natural language
- URL: http://arxiv.org/abs/2004.13839v2
- Date: Wed, 6 May 2020 10:30:24 GMT
- Title: Neural translation and automated recognition of ICD10 medical entities
from natural language
- Authors: Louis Falissard, Claire Morgand, Sylvie Roussel, Claire Imbaud, Walid
Ghosn, Karim Bounebache, Gr\'egoire Rey
- Abstract summary: The recognition of medical entities from natural language is an ubiquitous problem in the medical field.
The recent advances in artificial intelligence, specifically the raise of deep learning methods, has enabled computers to make efficient decisions.
This article investigates the applications of deep neural sequence models to the medical entity recognition from natural language problem.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recognition of medical entities from natural language is an ubiquitous
problem in the medical field, with applications ranging from medical act coding
to the analysis of electronic health data for public health. It is however a
complex task usually requiring human expert intervention, thus making it
expansive and time consuming. The recent advances in artificial intelligence,
specifically the raise of deep learning methods, has enabled computers to make
efficient decisions on a number of complex problems, with the notable example
of neural sequence models and their powerful applications in natural language
processing. They however require a considerable amount of data to learn from,
which is typically their main limiting factor. However, the C\'epiDc stores an
exhaustive database of death certificates at the French national scale,
amounting to several millions of natural language examples provided with their
associated human coded medical entities available to the machine learning
practitioner. This article investigates the applications of deep neural
sequence models to the medical entity recognition from natural language
problem.
Related papers
- Diagnostic Reasoning in Natural Language: Computational Model and Application [68.47402386668846]
We investigate diagnostic abductive reasoning (DAR) in the context of language-grounded tasks (NL-DAR)
We propose a novel modeling framework for NL-DAR based on Pearl's structural causal models.
We use the resulting dataset to investigate the human decision-making process in NL-DAR.
arXiv Detail & Related papers (2024-09-09T06:55:37Z) - Scientific Computing with Large Language Models [0.0]
We provide an overview of the emergence of large language models for scientific computing applications.
We highlight use cases that involve natural language processing of scientific documents and specialized languages designed to describe physical systems.
arXiv Detail & Related papers (2024-06-11T13:39:07Z) - Medical Vision-Language Pre-Training for Brain Abnormalities [96.1408455065347]
We show how to automatically collect medical image-text aligned data for pretraining from public resources such as PubMed.
In particular, we present a pipeline that streamlines the pre-training process by initially collecting a large brain image-text dataset.
We also investigate the unique challenge of mapping subfigures to subcaptions in the medical domain.
arXiv Detail & Related papers (2024-04-27T05:03:42Z) - Large language models in bioinformatics: applications and perspectives [14.16418711188321]
Large language models (LLMs) are artificial intelligence models based on deep learning.
This review focuses on exploring the applications of large language models in genomics, transcriptomics, drug discovery and single cell analysis.
arXiv Detail & Related papers (2024-01-08T17:26:59Z) - Language Generation from Brain Recordings [68.97414452707103]
We propose a generative language BCI that utilizes the capacity of a large language model and a semantic brain decoder.
The proposed model can generate coherent language sequences aligned with the semantic content of visual or auditory language stimuli.
Our findings demonstrate the potential and feasibility of employing BCIs in direct language generation.
arXiv Detail & Related papers (2023-11-16T13:37:21Z) - Advancing Italian Biomedical Information Extraction with
Transformers-based Models: Methodological Insights and Multicenter Practical
Application [0.27027468002793437]
Information Extraction can help clinical practitioners overcome the limitation by using automated text-mining pipelines.
We created the first Italian neuropsychiatric Named Entity Recognition dataset, PsyNIT, and used it to develop a Transformers-based model.
The lessons learned are: (i) the crucial role of a consistent annotation process and (ii) a fine-tuning strategy that combines classical methods with a "low-resource" approach.
arXiv Detail & Related papers (2023-06-08T16:15:46Z) - Language Models sounds the Death Knell of Knowledge Graphs [0.0]
Deep Learning based NLP especially Large Language Models (LLMs) have found broad acceptance and are used extensively for many applications.
BioBERT and Med-BERT are language models pre-trained for the healthcare domain.
This paper argues that using Knowledge Graphs is not the best solution for solving problems in this domain.
arXiv Detail & Related papers (2023-01-10T14:20:15Z) - Towards more patient friendly clinical notes through language models and
ontologies [57.51898902864543]
We present a novel approach to automated medical text based on word simplification and language modelling.
We use a new dataset pairs of publicly available medical sentences and a version of them simplified by clinicians.
Our method based on a language model trained on medical forum data generates simpler sentences while preserving both grammar and the original meaning.
arXiv Detail & Related papers (2021-12-23T16:11:19Z) - Neural Natural Language Processing for Unstructured Data in Electronic
Health Records: a Review [4.454501609622817]
Well over half of the information stored within EHRs is in the form of unstructured text.
Deep learning approaches to Natural Language Processing have made considerable advances.
We focus on a broad scope of tasks, namely, classification and prediction, word embeddings, extraction, generation, and other topics.
arXiv Detail & Related papers (2021-07-07T01:50:02Z) - CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark [51.38557174322772]
We present the first Chinese Biomedical Language Understanding Evaluation benchmark.
It is a collection of natural language understanding tasks including named entity recognition, information extraction, clinical diagnosis normalization, single-sentence/sentence-pair classification.
We report empirical results with the current 11 pre-trained Chinese models, and experimental results show that state-of-the-art neural models perform by far worse than the human ceiling.
arXiv Detail & Related papers (2021-06-15T12:25:30Z) - Domain-Specific Language Model Pretraining for Biomedical Natural
Language Processing [73.37262264915739]
We show that for domains with abundant unlabeled text, such as biomedicine, pretraining language models from scratch results in substantial gains.
Our experiments show that domain-specific pretraining serves as a solid foundation for a wide range of biomedical NLP tasks.
arXiv Detail & Related papers (2020-07-31T00:04:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.