Medical Concept Normalization in User Generated Texts by Learning Target
Concept Embeddings
- URL: http://arxiv.org/abs/2006.04014v1
- Date: Sun, 7 Jun 2020 01:17:18 GMT
- Title: Medical Concept Normalization in User Generated Texts by Learning Target
Concept Embeddings
- Authors: Katikapalli Subramanyam Kalyan, S.Sangeetha
- Abstract summary: Recent research approach concept normalization as either text classification or text matching.
Our proposed model overcomes these drawbacks by jointly learning the representations of input concept mention and target concepts.
Our model surpasses all the existing methods across three standard datasets by improving accuracy up to 2.31%.
- Score: 5.33024001730262
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical concept normalization helps in discovering standard concepts in
free-form text i.e., maps health-related mentions to standard concepts in a
vocabulary. It is much beyond simple string matching and requires a deep
semantic understanding of concept mentions. Recent research approach concept
normalization as either text classification or text matching. The main drawback
in existing a) text classification approaches is ignoring valuable target
concepts information in learning input concept mention representation b) text
matching approach is the need to separately generate target concept embeddings
which is time and resource consuming. Our proposed model overcomes these
drawbacks by jointly learning the representations of input concept mention and
target concepts. First, it learns the input concept mention representation
using RoBERTa. Second, it finds cosine similarity between embeddings of input
concept mention and all the target concepts. Here, embeddings of target
concepts are randomly initialized and then updated during training. Finally,
the target concept with maximum cosine similarity is assigned to the input
concept mention. Our model surpasses all the existing methods across three
standard datasets by improving accuracy up to 2.31%.
Related papers
- Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery [52.498055901649025]
Concept Bottleneck Models (CBMs) have been proposed to address the 'black-box' problem of deep neural networks.
We propose a novel CBM approach -- called Discover-then-Name-CBM (DN-CBM) -- that inverts the typical paradigm.
Our concept extraction strategy is efficient, since it is agnostic to the downstream task, and uses concepts already known to the model.
arXiv Detail & Related papers (2024-07-19T17:50:11Z) - Non-confusing Generation of Customized Concepts in Diffusion Models [135.4385383284657]
We tackle the common challenge of inter-concept visual confusion in compositional concept generation using text-guided diffusion models (TGDMs)
Existing customized generation methods only focus on fine-tuning the second stage while overlooking the first one.
We propose a simple yet effective solution called CLIF: contrastive image-language fine-tuning.
arXiv Detail & Related papers (2024-05-11T05:01:53Z) - Knowledge graphs for empirical concept retrieval [1.06378109904813]
Concept-based explainable AI is promising as a tool to improve the understanding of complex models at the premises of a given user.
Here, we present a workflow for user-driven data collection in both text and image domains.
We test the retrieved concept datasets on two concept-based explainability methods, namely concept activation vectors (CAVs) and concept activation regions (CARs)
arXiv Detail & Related papers (2024-04-10T13:47:22Z) - Can we Constrain Concept Bottleneck Models to Learn Semantically
Meaningful Input Features? [0.6993232019625149]
Concept Bottleneck Models (CBMs) are considered inherently interpretable because they first predict a set of human-defined concepts.
For inherent interpretability to be fully realised, we need to guarantee concepts are predicted based on semantically mapped input features.
We demonstrate that CBMs can learn concept representations with semantic mapping to input features by removing problematic concept correlations.
arXiv Detail & Related papers (2024-02-01T10:18:43Z) - Simple Mechanisms for Representing, Indexing and Manipulating Concepts [46.715152257557804]
We will argue that learning a concept could be done by looking at its moment statistics matrix to generate a concrete representation or signature of that concept.
When the concepts are intersected', signatures of the concepts can be used to find a common theme across a number of related intersected' concepts.
arXiv Detail & Related papers (2023-10-18T17:54:29Z) - Implicit Concept Removal of Diffusion Models [92.55152501707995]
Text-to-image (T2I) diffusion models often inadvertently generate unwanted concepts such as watermarks and unsafe images.
We present the Geom-Erasing, a novel concept removal method based on the geometric-driven control.
arXiv Detail & Related papers (2023-10-09T17:13:10Z) - LEACE: Perfect linear concept erasure in closed form [103.61624393221447]
Concept erasure aims to remove specified features from a representation.
We introduce LEAst-squares Concept Erasure (LEACE), a closed-form method which provably prevents all linear classifiers from detecting a concept while changing the representation as little as possible.
We apply LEACE to large language models with a novel procedure called "concept scrubbing," which erases target concept information from every layer in the network.
arXiv Detail & Related papers (2023-06-06T16:07:24Z) - DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for
Open-world Detection [118.36746273425354]
This paper presents a paralleled visual-concept pre-training method for open-world detection by resorting to knowledge enrichment from a designed concept dictionary.
By enriching the concepts with their descriptions, we explicitly build the relationships among various concepts to facilitate the open-domain learning.
The proposed framework demonstrates strong zero-shot detection performances, e.g., on the LVIS dataset, our DetCLIP-T outperforms GLIP-T by 9.9% mAP and obtains a 13.5% improvement on rare categories.
arXiv Detail & Related papers (2022-09-20T02:01:01Z) - FALCON: Fast Visual Concept Learning by Integrating Images, Linguistic
descriptions, and Conceptual Relations [99.54048050189971]
We present a framework for learning new visual concepts quickly, guided by multiple naturally occurring data streams.
The learned concepts support downstream applications, such as answering questions by reasoning about unseen images.
We demonstrate the effectiveness of our model on both synthetic and real-world datasets.
arXiv Detail & Related papers (2022-03-30T19:45:00Z) - A Context-based Disambiguation Model for Sentiment Concepts Using a
Bag-of-concepts Approach [0.0]
This study presents a context-based model to solve ambiguous polarity concepts using commonsense knowledge.
The proposed model is evaluated by applying a corpus of product reviews, called Semeval.
The experimental results revealed an accuracy rate of 82.07%, representing the effectiveness of the proposed model.
arXiv Detail & Related papers (2020-08-07T07:16:40Z) - On the Learnability of Concepts: With Applications to Comparing Word
Embedding Algorithms [0.0]
We introduce the notion of "concept" as a list of words that have shared semantic content.
We first use this notion to measure the learnability of concepts on pretrained word embeddings.
We then develop a statistical analysis of concept learnability, based on hypothesis testing and ROC curves, in order to compare the relative merits of various embedding algorithms.
arXiv Detail & Related papers (2020-06-17T14:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.