Incorporating Dictionaries into a Neural Network Architecture to Extract
COVID-19 Medical Concepts From Social Media
- URL: http://arxiv.org/abs/2309.02188v1
- Date: Tue, 5 Sep 2023 12:47:44 GMT
- Title: Incorporating Dictionaries into a Neural Network Architecture to Extract
COVID-19 Medical Concepts From Social Media
- Authors: Abul Hasan and Mark Levene and David Weston
- Abstract summary: We investigate the potential benefit of incorporating dictionary information into a neural network architecture for natural language processing.
In particular, we make use of this architecture to extract several concepts related to COVID-19 from an on-line medical forum.
Our results show that incorporating small domain dictionaries to deep learning models can improve concept extraction tasks.
- Score: 0.2302001830524133
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We investigate the potential benefit of incorporating dictionary information
into a neural network architecture for natural language processing. In
particular, we make use of this architecture to extract several concepts
related to COVID-19 from an on-line medical forum. We use a sample from the
forum to manually curate one dictionary for each concept. In addition, we use
MetaMap, which is a tool for extracting biomedical concepts, to identify a
small number of semantic concepts. For a supervised concept extraction task on
the forum data, our best model achieved a macro $F_1$ score of 90\%. A major
difficulty in medical concept extraction is obtaining labelled data from which
to build supervised models. We investigate the utility of our models to
transfer to data derived from a different source in two ways. First for
producing labels via weak learning and second to perform concept extraction.
The dataset we use in this case comprises COVID-19 related tweets and we
achieve an $F_1$ score 81\% for symptom concept extraction trained on weakly
labelled data. The utility of our dictionaries is compared with a COVID-19
symptom dictionary that was constructed directly from Twitter. Further
experiments that incorporate BERT and a COVID-19 version of BERTweet
demonstrate that the dictionaries provide a commensurate result. Our results
show that incorporating small domain dictionaries to deep learning models can
improve concept extraction tasks. Moreover, models built using dictionaries
generalize well and are transferable to different datasets on a similar task.
Related papers
- Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery [52.498055901649025]
Concept Bottleneck Models (CBMs) have been proposed to address the 'black-box' problem of deep neural networks.
We propose a novel CBM approach -- called Discover-then-Name-CBM (DN-CBM) -- that inverts the typical paradigm.
Our concept extraction strategy is efficient, since it is agnostic to the downstream task, and uses concepts already known to the model.
arXiv Detail & Related papers (2024-07-19T17:50:11Z) - DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for
Open-world Detection [118.36746273425354]
This paper presents a paralleled visual-concept pre-training method for open-world detection by resorting to knowledge enrichment from a designed concept dictionary.
By enriching the concepts with their descriptions, we explicitly build the relationships among various concepts to facilitate the open-domain learning.
The proposed framework demonstrates strong zero-shot detection performances, e.g., on the LVIS dataset, our DetCLIP-T outperforms GLIP-T by 9.9% mAP and obtains a 13.5% improvement on rare categories.
arXiv Detail & Related papers (2022-09-20T02:01:01Z) - A Unified Model for Reverse Dictionary and Definition Modelling [7.353994554197792]
We train a dual-way neural dictionary to guess words from definitions (reverse dictionary) and produce definitions given words (definition modelling)
Our method learns the two tasks simultaneously, and handles unknown words via embeddings.
It casts a word or a definition to the same representation space through a shared layer, then generates the other form from there, in a multi-task fashion.
arXiv Detail & Related papers (2022-05-09T23:52:39Z) - Meta-Learning with Variational Semantic Memory for Word Sense
Disambiguation [56.830395467247016]
We propose a model of semantic memory for WSD in a meta-learning setting.
Our model is based on hierarchical variational inference and incorporates an adaptive memory update rule via a hypernetwork.
We show our model advances the state of the art in few-shot WSD, supports effective learning in extremely data scarce scenarios.
arXiv Detail & Related papers (2021-06-05T20:40:01Z) - End-to-end Biomedical Entity Linking with Span-based Dictionary Matching [5.273138059454523]
Disease name recognition and normalization is a fundamental process in biomedical text mining.
This study introduces a novel end-to-end approach that combines span representations with dictionary-matching features.
Our model handles unseen concepts by referring to a dictionary while maintaining the performance of neural network-based models.
arXiv Detail & Related papers (2021-04-21T12:24:12Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - When Dictionary Learning Meets Deep Learning: Deep Dictionary Learning
and Coding Network for Image Recognition with Limited Data [74.75557280245643]
We present a new Deep Dictionary Learning and Coding Network (DDLCN) for image recognition tasks with limited data.
We empirically compare DDLCN with several leading dictionary learning methods and deep learning models.
Experimental results on five popular datasets show that DDLCN achieves competitive results compared with state-of-the-art methods when the training data is limited.
arXiv Detail & Related papers (2020-05-21T23:12:10Z) - 3D medical image segmentation with labeled and unlabeled data using
autoencoders at the example of liver segmentation in CT images [58.720142291102135]
This work investigates the potential of autoencoder-extracted features to improve segmentation with a convolutional neural network.
A convolutional autoencoder was used to extract features from unlabeled data and a multi-scale, fully convolutional CNN was used to perform the target task of 3D liver segmentation in CT images.
arXiv Detail & Related papers (2020-03-17T20:20:43Z) - Learning Deep Analysis Dictionaries for Image Super-Resolution [38.7315182732103]
Deep Analysis dictionary Model (DeepAM) is optimized to address a specific regression task known as single image super-resolution.
Our architecture contains L layers of analysis dictionary and soft-thresholding operators.
DeepAM uses both supervised and unsupervised setup.
arXiv Detail & Related papers (2020-01-31T18:59:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.