Domain Adaptative Causality Encoder
- URL: http://arxiv.org/abs/2011.13549v1
- Date: Fri, 27 Nov 2020 04:14:55 GMT
- Title: Domain Adaptative Causality Encoder
- Authors: Farhad Moghimifar, Gholamreza Haffari, Mahsa Baktashmotlagh
- Abstract summary: We leverage the characteristics of dependency trees and adversarial learning to address the tasks of adaptive causality identification and localisation.
We present a new causality dataset, namely MedCaus, which integrates all types of causality in the text.
- Score: 52.779274858332656
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current approaches which are mainly based on the extraction of low-level
relations among individual events are limited by the shortage of publicly
available labelled data. Therefore, the resulting models perform poorly when
applied to a distributionally different domain for which labelled data did not
exist at the time of training. To overcome this limitation, in this paper, we
leverage the characteristics of dependency trees and adversarial learning to
address the tasks of adaptive causality identification and localisation. The
term adaptive is used since the training and test data come from two
distributionally different datasets, which to the best of our knowledge, this
work is the first to address. Moreover, we present a new causality dataset,
namely MedCaus, which integrates all types of causality in the text. Our
experiments on four different benchmark causality datasets demonstrate the
superiority of our approach over the existing baselines, by up to 7%
improvement, on the tasks of identification and localisation of the causal
relations from the text.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - Autoencoder based approach for the mitigation of spurious correlations [2.7624021966289605]
Spurious correlations refer to erroneous associations in data that do not reflect true underlying relationships.
These correlations can lead deep neural networks (DNNs) to learn patterns that are not robust across diverse datasets or real-world scenarios.
We propose an autoencoder-based approach to analyze the nature of spurious correlations that exist in the Global Wheat Head Detection (GWHD) 2021 dataset.
arXiv Detail & Related papers (2024-06-27T05:28:44Z) - Disentangle Estimation of Causal Effects from Cross-Silo Data [14.684584362172666]
We introduce an innovative disentangle architecture designed to facilitate the seamless cross-silo transmission of model parameters.
We introduce global constraints into the equation to effectively mitigate bias within the various missing domains.
Our method outperforms state-of-the-art baselines.
arXiv Detail & Related papers (2024-01-04T09:05:37Z) - Holistic Transfer: Towards Non-Disruptive Fine-Tuning with Partial
Target Data [32.91362206231936]
We propose a learning problem involving adapting a pre-trained source model to the target domain for classifying all classes that appeared in the source data.
This problem is practical, as it is unrealistic for the target end-users to collect data for all classes prior to adaptation.
We present several effective solutions that maintain the accuracy of the missing classes and enhance the overall performance.
arXiv Detail & Related papers (2023-11-02T17:35:16Z) - SALUDA: Surface-based Automotive Lidar Unsupervised Domain Adaptation [62.889835139583965]
We introduce an unsupervised auxiliary task of learning an implicit underlying surface representation simultaneously on source and target data.
As both domains share the same latent representation, the model is forced to accommodate discrepancies between the two sources of data.
Our experiments demonstrate that our method achieves a better performance than the current state of the art, both in real-to-real and synthetic-to-real scenarios.
arXiv Detail & Related papers (2023-04-06T17:36:23Z) - Direct-Effect Risk Minimization for Domain Generalization [11.105832297850188]
We introduce the concepts of direct and indirect effects from causal inference to the domain generalization problem.
We argue that models that learn direct effects minimize the worst-case risk across correlation-shifted domains.
Experiments on 5 correlation-shifted datasets and the DomainBed benchmark verify the effectiveness of our approach.
arXiv Detail & Related papers (2022-11-26T15:35:36Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z) - Discriminative Active Learning for Domain Adaptation [16.004653151961303]
We introduce a discriminative active learning approach for domain adaptation to reduce the efforts of data annotation.
Specifically, we propose three-stage active adversarial training of neural networks.
Empirical comparisons with existing domain adaptation methods using four benchmark datasets demonstrate the effectiveness of the proposed approach.
arXiv Detail & Related papers (2020-05-24T04:20:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.