Causal Feature Selection with Dimension Reduction for Interpretable Text
Classification
- URL: http://arxiv.org/abs/2010.04609v1
- Date: Fri, 9 Oct 2020 14:36:49 GMT
- Title: Causal Feature Selection with Dimension Reduction for Interpretable Text
Classification
- Authors: Guohou Shan, James Foulds, Shimei Pan
- Abstract summary: We investigate a class of matching-based causal inference methods for text feature selection.
We propose a new causal feature selection framework that combines dimension reduction with causal inference to improve text feature selection.
- Score: 7.20833506531457
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text features that are correlated with class labels, but do not directly
cause them, are sometimesuseful for prediction, but they may not be insightful.
As an alternative to traditional correlation-basedfeature selection, causal
inference could reveal more principled, meaningful relationships betweentext
features and labels. To help researchers gain insight into text data, e.g. for
social scienceapplications, in this paper we investigate a class of
matching-based causal inference methods fortext feature selection. Features
used in document classification are often high dimensional, howeverexisting
causal feature selection methods use Propensity Score Matching (PSM) which is
known to beless effective in high-dimensional spaces. We propose a new causal
feature selection framework thatcombines dimension reduction with causal
inference to improve text feature selection. Experiments onboth synthetic and
real-world data demonstrate the promise of our methods in improving
classificationand enhancing interpretability.
Related papers
- CAST: Corpus-Aware Self-similarity Enhanced Topic modelling [16.562349140796115]
We introduce CAST: Corpus-Aware Self-similarity Enhanced Topic modelling, a novel topic modelling method.
We find self-similarity to be an effective metric to prevent functional words from acting as candidate topic words.
Our approach significantly enhances the coherence and diversity of generated topics, as well as the topic model's ability to handle noisy data.
arXiv Detail & Related papers (2024-10-19T15:27:11Z) - Causal Feature Selection via Transfer Entropy [59.999594949050596]
Causal discovery aims to identify causal relationships between features with observational data.
We introduce a new causal feature selection approach that relies on the forward and backward feature selection procedures.
We provide theoretical guarantees on the regression and classification errors for both the exact and the finite-sample cases.
arXiv Detail & Related papers (2023-10-17T08:04:45Z) - IDEAL: Influence-Driven Selective Annotations Empower In-Context
Learners in Large Language Models [66.32043210237768]
This paper introduces an influence-driven selective annotation method.
It aims to minimize annotation costs while improving the quality of in-context examples.
Experiments confirm the superiority of the proposed method on various benchmarks.
arXiv Detail & Related papers (2023-10-16T22:53:54Z) - Description-Based Text Similarity [59.552704474862004]
We identify the need to search for texts based on abstract descriptions of their content.
We propose an alternative model that significantly improves when used in standard nearest neighbor search.
arXiv Detail & Related papers (2023-05-21T17:14:31Z) - Selective Text Augmentation with Word Roles for Low-Resource Text
Classification [3.4806267677524896]
Different words may play different roles in text classification, which inspires us to strategically select the proper roles for text augmentation.
In this work, we first identify the relationships between the words in a text and the text category from the perspectives of statistical correlation and semantic similarity.
We present a new augmentation technique called STA (Selective Text Augmentation) where different text-editing operations are selectively applied to words with specific roles.
arXiv Detail & Related papers (2022-09-04T08:13:11Z) - Hierarchical Heterogeneous Graph Representation Learning for Short Text
Classification [60.233529926965836]
We propose a new method called SHINE, which is based on graph neural network (GNN) for short text classification.
First, we model the short text dataset as a hierarchical heterogeneous graph consisting of word-level component graphs.
Then, we dynamically learn a short document graph that facilitates effective label propagation among similar short texts.
arXiv Detail & Related papers (2021-10-30T05:33:05Z) - Does a Hybrid Neural Network based Feature Selection Model Improve Text
Classification? [9.23545668304066]
We propose a hybrid feature selection method for obtaining relevant features.
We then present three ways of implementing a feature selection and neural network pipeline.
We also observed a slight increase in accuracy on some datasets.
arXiv Detail & Related papers (2021-01-22T09:12:19Z) - Classifying Scientific Publications with BERT -- Is Self-Attention a
Feature Selection Method? [0.0]
We investigate the self-attention mechanism of BERT in a fine-tuning scenario for the classification of scientific articles.
We observe how self-attention focuses on words that are highly related to the domain of the article.
We compare and evaluate the subset of the most attended words with feature selection methods normally used for text classification.
arXiv Detail & Related papers (2021-01-20T13:22:26Z) - Weakly-Supervised Aspect-Based Sentiment Analysis via Joint
Aspect-Sentiment Topic Embedding [71.2260967797055]
We propose a weakly-supervised approach for aspect-based sentiment analysis.
We learn sentiment, aspect> joint topic embeddings in the word embedding space.
We then use neural models to generalize the word-level discriminative information.
arXiv Detail & Related papers (2020-10-13T21:33:24Z) - Identifying Spurious Correlations for Robust Text Classification [9.457737910527829]
We propose a method to distinguish spurious and genuine correlations in text classification.
We use features derived from treatment effect estimators to distinguish spurious correlations from "genuine" ones.
Experiments on four datasets suggest that using this approach to inform feature selection also leads to more robust classification.
arXiv Detail & Related papers (2020-10-06T03:49:22Z) - Don't Judge an Object by Its Context: Learning to Overcome Contextual
Bias [113.44471186752018]
Existing models often leverage co-occurrences between objects and their context to improve recognition accuracy.
This work focuses on addressing such contextual biases to improve the robustness of the learnt feature representations.
arXiv Detail & Related papers (2020-01-09T18:31:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.