Causally Estimating the Sensitivity of Neural NLP Models to Spurious
Features
- URL: http://arxiv.org/abs/2110.07159v1
- Date: Thu, 14 Oct 2021 05:26:08 GMT
- Title: Causally Estimating the Sensitivity of Neural NLP Models to Spurious
Features
- Authors: Yunxiang Zhang, Liangming Pan, Samson Tan, Min-Yen Kan
- Abstract summary: There is no measure to evaluate or compare the effects of different forms of spurious features in NLP.
We quantify model sensitivity to spurious features with a causal estimand, dubbed CENT.
We find statistically significant inverse correlations between sensitivity and robustness, providing empirical support for our hypothesis.
- Score: 19.770032728328733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work finds modern natural language processing (NLP) models relying on
spurious features for prediction. Mitigating such effects is thus important.
Despite this need, there is no quantitative measure to evaluate or compare the
effects of different forms of spurious features in NLP. We address this gap in
the literature by quantifying model sensitivity to spurious features with a
causal estimand, dubbed CENT, which draws on the concept of average treatment
effect from the causality literature. By conducting simulations with four
prominent NLP models -- TextRNN, BERT, RoBERTa and XLNet -- we rank the models
against their sensitivity to artificial injections of eight spurious features.
We further hypothesize and validate that models that are more sensitive to a
spurious feature will be less robust against perturbations with this feature
during inference. Conversely, data augmentation with this feature improves
robustness to similar perturbations. We find statistically significant inverse
correlations between sensitivity and robustness, providing empirical support
for our hypothesis.
Related papers
- The Impact of Quantization on the Robustness of Transformer-based Text
Classifiers [5.281054432963503]
This work is the first application of quantization on the robustness of NLP models.
We evaluate the impact of quantization on BERT and DistilBERT models in text classification using SST-2, Emotion, and MR datasets.
Our experiments indicate that quantization increases the robustness of the model by 18.80% on average compared to adversarial training.
arXiv Detail & Related papers (2024-03-08T14:55:05Z) - The Risk of Federated Learning to Skew Fine-Tuning Features and
Underperform Out-of-Distribution Robustness [50.52507648690234]
Federated learning has the risk of skewing fine-tuning features and compromising the robustness of the model.
We introduce three robustness indicators and conduct experiments across diverse robust datasets.
Our approach markedly enhances the robustness across diverse scenarios, encompassing various parameter-efficient fine-tuning methods.
arXiv Detail & Related papers (2024-01-25T09:18:51Z) - Beyond the Universal Law of Robustness: Sharper Laws for Random Features
and Neural Tangent Kernels [14.186776881154127]
This paper focuses on empirical risk minimization in two settings, namely, random features and the neural tangent kernel (NTK)
We prove that, for random features, the model is not robust for any degree of over- parameterization, even when the necessary condition coming from the universal law of robustness is satisfied.
Our results are corroborated by numerical evidence on both synthetic and standard prototypical datasets.
arXiv Detail & Related papers (2023-02-03T09:58:31Z) - Improving Prediction of Cognitive Performance using Deep Neural Networks
in Sparse Data [2.867517731896504]
We used data from an observational, cohort study, Midlife in the United States (MIDUS) to model executive function and episodic memory measures.
Deep neural network (DNN) models consistently ranked highest in all of the cognitive performance prediction tasks.
arXiv Detail & Related papers (2021-12-28T22:23:08Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Balancing Robustness and Sensitivity using Feature Contrastive Learning [95.86909855412601]
Methods that promote robustness can hurt the model's sensitivity to rare or underrepresented patterns.
We propose Feature Contrastive Learning (FCL) that encourages a model to be more sensitive to the features that have higher contextual utility.
arXiv Detail & Related papers (2021-05-19T20:53:02Z) - Efficient Causal Inference from Combined Observational and
Interventional Data through Causal Reductions [68.6505592770171]
Unobserved confounding is one of the main challenges when estimating causal effects.
We propose a novel causal reduction method that replaces an arbitrary number of possibly high-dimensional latent confounders.
We propose a learning algorithm to estimate the parameterized reduced model jointly from observational and interventional data.
arXiv Detail & Related papers (2021-03-08T14:29:07Z) - Non-Singular Adversarial Robustness of Neural Networks [58.731070632586594]
Adrial robustness has become an emerging challenge for neural network owing to its over-sensitivity to small input perturbations.
We formalize the notion of non-singular adversarial robustness for neural networks through the lens of joint perturbations to data inputs as well as model weights.
arXiv Detail & Related papers (2021-02-23T20:59:30Z) - Removing Spurious Features can Hurt Accuracy and Affect Groups
Disproportionately [83.68135652247496]
A natural remedy is to remove spurious features from the model.
We show that removal of spurious features can decrease accuracy due to inductive biases.
We also show that robust self-training can remove spurious features without affecting the overall accuracy.
arXiv Detail & Related papers (2020-12-07T23:08:59Z) - One Neuron to Fool Them All [12.107259467873094]
We evaluate the sensitivity of individual neurons in terms of how robust the model's output is to direct perturbations of that neuron's output.
Attacks using a loss function that targets just a single sensitive neuron find adversarial examples nearly as effectively as ones that target the full model.
arXiv Detail & Related papers (2020-03-20T16:49:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.