CERM: Context-aware Literature-based Discovery via Sentiment Analysis
- URL: http://arxiv.org/abs/2402.01724v1
- Date: Sat, 27 Jan 2024 06:40:08 GMT
- Title: CERM: Context-aware Literature-based Discovery via Sentiment Analysis
- Authors: Julio Christian Young and Uchenna Akujuobi
- Abstract summary: This paper introduces Entity Relationship Sentiment Analysis (ERSA), a new task that captures the sentiment of a text based on an entity pair.
ERSA poses a significant challenge compared to traditional sentiment analysis tasks, as sentence sentiment may not align with entity relationship sentiment.
We propose CERM, a semi-supervised architecture that combines different word embeddings to enhance the encoding of the ERSA task.
- Score: 1.6317061277457001
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Driven by the abundance of biomedical publications, we introduce a sentiment
analysis task to understand food-health relationship. Prior attempts to
incorporate health into recipe recommendation and analysis systems have
primarily focused on ingredient nutritional components or utilized basic
computational models trained on curated labeled data. Enhanced models that
capture the inherent relationship between food ingredients and biomedical
concepts can be more beneficial for food-related research, given the wealth of
information in biomedical texts. Considering the costly data labeling process,
these models should effectively utilize both labeled and unlabeled data. This
paper introduces Entity Relationship Sentiment Analysis (ERSA), a new task that
captures the sentiment of a text based on an entity pair. ERSA extends the
widely studied Aspect Based Sentiment Analysis (ABSA) task. Specifically, our
study concentrates on the ERSA task applied to biomedical texts, focusing on
(entity-entity) pairs of biomedical and food concepts. ERSA poses a significant
challenge compared to traditional sentiment analysis tasks, as sentence
sentiment may not align with entity relationship sentiment. Additionally, we
propose CERM, a semi-supervised architecture that combines different word
embeddings to enhance the encoding of the ERSA task. Experimental results
showcase the model's efficiency across diverse learning scenarios.
Related papers
- Causal Discovery from Data Assisted by Large Language Models [50.193740129296245]
It is essential to integrate experimental data with prior domain knowledge for knowledge driven discovery.
Here we demonstrate this approach by combining high-resolution scanning transmission electron microscopy (STEM) data with insights derived from large language models (LLMs)
By fine-tuning ChatGPT on domain-specific literature, we construct adjacency matrices for Directed Acyclic Graphs (DAGs) that map the causal relationships between structural, chemical, and polarization degrees of freedom in Sm-doped BiFeO3 (SmBFO)
arXiv Detail & Related papers (2025-03-18T02:14:49Z) - Enhancing Biomedical Relation Extraction with Directionality [4.0241840878351764]
We propose a novel multi-task language model with soft-prompt learning to jointly identify the relationship, novel findings, and entity roles.
Our results in-clude an enriched BioRED corpus with 10,864 directionality annotations.
arXiv Detail & Related papers (2025-01-23T20:36:11Z) - Biomedical Relation Extraction via Adaptive Document-Relation Cross-Mapping and Concept Unique Identifier [35.79876359248485]
Document-Level Biomedical Relation Extraction (Bio-RE) aims to identify relations between biomedical entities within extensive texts.
Previous methods often overlook the incompleteness of documents and lack the integration of external knowledge.
Recent advancements in large language models (LLMs) have inspired us to explore all the above issues for document-level Bio-RE.
arXiv Detail & Related papers (2025-01-09T11:19:40Z) - Causal Representation Learning from Multimodal Biological Observations [57.00712157758845]
We aim to develop flexible identification conditions for multimodal data.
We establish identifiability guarantees for each latent component, extending the subspace identification results from prior work.
Our key theoretical ingredient is the structural sparsity of the causal connections among distinct modalities.
arXiv Detail & Related papers (2024-11-10T16:40:27Z) - Extracting Protein-Protein Interactions (PPIs) from Biomedical
Literature using Attention-based Relational Context Information [5.456047952635665]
This work presents a unified, multi-source PPI corpora with vetted interaction definitions augmented by binary interaction type labels.
A Transformer-based deep learning method exploits entities' relational context information for relation representation to improve relation classification performance.
The model's performance is evaluated on four widely studied biomedical relation extraction datasets.
arXiv Detail & Related papers (2024-03-08T01:43:21Z) - Diversifying Knowledge Enhancement of Biomedical Language Models using
Adapter Modules and Knowledge Graphs [54.223394825528665]
We develop an approach that uses lightweight adapter modules to inject structured biomedical knowledge into pre-trained language models.
We use two large KGs, the biomedical knowledge system UMLS and the novel biochemical OntoChem, with two prominent biomedical PLMs, PubMedBERT and BioLinkBERT.
We show that our methodology leads to performance improvements in several instances while keeping requirements in computing power low.
arXiv Detail & Related papers (2023-12-21T14:26:57Z) - High-throughput Biomedical Relation Extraction for Semi-Structured Web Articles Empowered by Large Language Models [1.9665865095034865]
We formulate the relation extraction task as binary classifications for large language models.
We designate the main title as the tail entity and explicitly incorporate it into the context.
Longer contents are sliced into text chunks, embedded, and retrieved with additional embedding models.
arXiv Detail & Related papers (2023-12-13T16:43:41Z) - EBOCA: Evidences for BiOmedical Concepts Association Ontology [55.41644538483948]
This paper proposes EBOCA, an ontology that describes (i) biomedical domain concepts and associations between them, and (ii) evidences supporting these associations.
Test data coming from a subset of DISNET and automatic association extractions from texts has been transformed to create a Knowledge Graph that can be used in real scenarios.
arXiv Detail & Related papers (2022-08-01T18:47:03Z) - BioIE: Biomedical Information Extraction with Multi-head Attention
Enhanced Graph Convolutional Network [9.227487525657901]
We propose Biomedical Information Extraction, a hybrid neural network to extract relations from biomedical text and unstructured medical reports.
We evaluate our model on two major biomedical relationship extraction tasks, chemical-disease relation and chemical-protein interaction, and a cross-hospital pan-cancer pathology report corpus.
arXiv Detail & Related papers (2021-10-26T13:19:28Z) - Discovering Drug-Target Interaction Knowledge from Biomedical Literature [107.98712673387031]
The Interaction between Drugs and Targets (DTI) in human body plays a crucial role in biomedical science and applications.
As millions of papers come out every year in the biomedical domain, automatically discovering DTI knowledge from literature becomes an urgent demand in the industry.
We explore the first end-to-end solution for this task by using generative approaches.
We regard the DTI triplets as a sequence and use a Transformer-based model to directly generate them without using the detailed annotations of entities and relations.
arXiv Detail & Related papers (2021-09-27T17:00:14Z) - MIMO: Mutual Integration of Patient Journey and Medical Ontology for
Healthcare Representation Learning [49.57261599776167]
We propose an end-to-end robust Transformer-based solution, Mutual Integration of patient journey and Medical Ontology (MIMO) for healthcare representation learning and predictive analytics.
arXiv Detail & Related papers (2021-07-20T07:04:52Z) - Scientific Language Models for Biomedical Knowledge Base Completion: An
Empirical Study [62.376800537374024]
We study scientific LMs for KG completion, exploring whether we can tap into their latent knowledge to enhance biomedical link prediction.
We integrate the LM-based models with KG embedding models, using a router method that learns to assign each input example to either type of model and provides a substantial boost in performance.
arXiv Detail & Related papers (2021-06-17T17:55:33Z) - PharmKE: Knowledge Extraction Platform for Pharmaceutical Texts using
Transfer Learning [0.0]
PharmKE is a text analysis platform that applies deep learning through several stages for thorough semantic analysis of pharmaceutical articles.
The methodology is used to create accurately labeled training and test datasets, which are then used to train models for custom entity labeling tasks.
The obtained results are compared to the fine-tuned BERT and BioBERT models trained on the same dataset.
arXiv Detail & Related papers (2021-02-25T19:36:35Z) - Chemical-induced Disease Relation Extraction with Dependency Information
and Prior Knowledge [2.9686294158279414]
We propose a novel convolutional attention network (CAN) for chemical-disease relation (CDR) extraction.
Firstly, we extract the shortest dependency path (SDP) between chemical and disease pairs in a sentence.
After that, an attention mechanism is employed to learn the importance/weight of each semantic dependency vector related to knowledge representations learned from KBs.
arXiv Detail & Related papers (2020-01-02T02:24:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.