Contextual information integration for stance detection via
cross-attention
- URL: http://arxiv.org/abs/2211.01874v2
- Date: Thu, 25 May 2023 12:47:37 GMT
- Title: Contextual information integration for stance detection via
cross-attention
- Authors: Tilman Beck, Andreas Waldis, Iryna Gurevych
- Abstract summary: Stance detection deals with identifying an author's stance towards a target.
Most existing stance detection models are limited because they do not consider relevant contextual information.
We propose an approach to integrate contextual information as text.
- Score: 59.662413798388485
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stance detection deals with identifying an author's stance towards a target.
Most existing stance detection models are limited because they do not consider
relevant contextual information which allows for inferring the stance
correctly. Complementary context can be found in knowledge bases but
integrating the context into pretrained language models is non-trivial due to
the graph structure of standard knowledge bases. To overcome this, we explore
an approach to integrate contextual information as text which allows for
integrating contextual information from heterogeneous sources, such as
structured knowledge sources and by prompting large language models. Our
approach can outperform competitive baselines on a large and diverse stance
detection benchmark in a cross-target setup, i.e. for targets unseen during
training. We demonstrate that it is more robust to noisy context and can
regularize for unwanted correlations between labels and target-specific
vocabulary. Finally, it is independent of the pretrained language model in use.
Related papers
- End-to-end Open-vocabulary Video Visual Relationship Detection using Multi-modal Prompting [68.37943632270505]
Open-vocabulary video visual relationship detection aims to expand video visual relationship detection beyond categories.
Existing methods usually use trajectory detectors trained on closed datasets to detect object trajectories.
We propose an open-vocabulary relationship that leverages the rich semantic knowledge of CLIP to discover novel relationships.
arXiv Detail & Related papers (2024-09-19T06:25:01Z) - Promoting Target Data in Context-aware Neural Machine Translation [1.8130068086063336]
We evaluate novel concatenation-based variants where the target context is prepended to the source language.
We show that including target context in the source leads to large improvements on target language phenomena.
arXiv Detail & Related papers (2024-02-09T11:34:39Z) - Quantifying the Plausibility of Context Reliance in Neural Machine
Translation [25.29330352252055]
We introduce Plausibility Evaluation of Context Reliance (PECoRe)
PECoRe is an end-to-end interpretability framework designed to quantify context usage in language models' generations.
We use pecore to quantify the plausibility of context-aware machine translation models.
arXiv Detail & Related papers (2023-10-02T13:26:43Z) - Guiding Computational Stance Detection with Expanded Stance Triangle
Framework [25.2980607215715]
Stance detection determines whether the author of a piece of text is in favor of, against, or neutral towards a specified target.
We decompose the stance detection task from a linguistic perspective, and investigate key components and inference paths in this task.
arXiv Detail & Related papers (2023-05-31T13:33:29Z) - Compositional Generalization in Grounded Language Learning via Induced
Model Sparsity [81.38804205212425]
We consider simple language-conditioned navigation problems in a grid world environment with disentangled observations.
We design an agent that encourages sparse correlations between words in the instruction and attributes of objects, composing them together to find the goal.
Our agent maintains a high level of performance on goals containing novel combinations of properties even when learning from a handful of demonstrations.
arXiv Detail & Related papers (2022-07-06T08:46:27Z) - Exploiting Multi-Object Relationships for Detecting Adversarial Attacks
in Complex Scenes [51.65308857232767]
Vision systems that deploy Deep Neural Networks (DNNs) are known to be vulnerable to adversarial examples.
Recent research has shown that checking the intrinsic consistencies in the input data is a promising way to detect adversarial attacks.
We develop a novel approach to perform context consistency checks using language models.
arXiv Detail & Related papers (2021-08-19T00:52:10Z) - Contextual Diversity for Active Learning [9.546771465714876]
Large datasets restrict the use of deep convolutional neural networks (CNNs) for many practical applications.
We introduce the notion of contextual diversity that captures the confusion associated with spatially co-occurring classes.
Our studies show clear advantages of using contextual diversity for active learning.
arXiv Detail & Related papers (2020-08-13T07:04:15Z) - How Context Affects Language Models' Factual Predictions [134.29166998377187]
We integrate information from a retrieval system with a pre-trained language model in a purely unsupervised way.
We report that augmenting pre-trained language models in this way dramatically improves performance and that the resulting system, despite being unsupervised, is competitive with a supervised machine reading baseline.
arXiv Detail & Related papers (2020-05-10T09:28:12Z) - Fast and Robust Unsupervised Contextual Biasing for Speech Recognition [16.557586847398778]
We propose an alternative approach that does not entail explicit contextual language model.
We derive the bias score for every word in the system vocabulary from the training corpus.
We show significant improvement in recognition accuracy when the relevant context is available.
arXiv Detail & Related papers (2020-05-04T17:29:59Z) - Exploiting Structured Knowledge in Text via Graph-Guided Representation
Learning [73.0598186896953]
We present two self-supervised tasks learning over raw text with the guidance from knowledge graphs.
Building upon entity-level masked language models, our first contribution is an entity masking scheme.
In contrast to existing paradigms, our approach uses knowledge graphs implicitly, only during pre-training.
arXiv Detail & Related papers (2020-04-29T14:22:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.