Deep Active Learning for Text Classification with Diverse
Interpretations
- URL: http://arxiv.org/abs/2108.10687v1
- Date: Sun, 15 Aug 2021 10:42:07 GMT
- Title: Deep Active Learning for Text Classification with Diverse
Interpretations
- Authors: Qiang Liu and Yanqiao Zhu and Zhaocheng Liu and Yufeng Zhang and Shu
Wu
- Abstract summary: We propose a novel Active Learning with DivErse iNterpretations (ALDEN) approach.
With local interpretations in Deep Neural Networks (DNNs), ALDEN identifies linearly separable regions of samples.
To tackle the text classification problem, we choose the word with the most diverse interpretations to represent the whole sentence.
- Score: 20.202134075256094
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, Deep Neural Networks (DNNs) have made remarkable progress for text
classification, which, however, still require a large number of labeled data.
To train high-performing models with the minimal annotation cost, active
learning is proposed to select and label the most informative samples, yet it
is still challenging to measure informativeness of samples used in DNNs. In
this paper, inspired by piece-wise linear interpretability of DNNs, we propose
a novel Active Learning with DivErse iNterpretations (ALDEN) approach. With
local interpretations in DNNs, ALDEN identifies linearly separable regions of
samples. Then, it selects samples according to their diversity of local
interpretations and queries their labels. To tackle the text classification
problem, we choose the word with the most diverse interpretations to represent
the whole sentence. Extensive experiments demonstrate that ALDEN consistently
outperforms several state-of-the-art deep active learning methods.
Related papers
- Beyond Coarse-Grained Matching in Video-Text Retrieval [50.799697216533914]
We introduce a new approach for fine-grained evaluation.
Our approach can be applied to existing datasets by automatically generating hard negative test captions.
Experiments on our fine-grained evaluations demonstrate that this approach enhances a model's ability to understand fine-grained differences.
arXiv Detail & Related papers (2024-10-16T09:42:29Z) - Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - SLCNN: Sentence-Level Convolutional Neural Network for Text
Classification [0.0]
Convolutional neural network (CNN) has shown remarkable success in the task of text classification.
New baseline models have been studied for text classification using CNN.
Results have shown that the proposed models have better performance, particularly in the longer documents.
arXiv Detail & Related papers (2023-01-27T13:16:02Z) - A semantic hierarchical graph neural network for text classification [1.439766998338892]
We propose a new hierarchical graph neural network (HieGNN) which extracts corresponding information from word-level, sentence-level and document-level respectively.
Experimental results on several benchmark datasets achieve better or similar results compared to several baseline methods.
arXiv Detail & Related papers (2022-09-15T03:59:31Z) - Learning to Imagine: Diversify Memory for Incremental Learning using
Unlabeled Data [69.30452751012568]
We develop a learnable feature generator to diversify exemplars by adaptively generating diverse counterparts of exemplars.
We introduce semantic contrastive learning to enforce the generated samples to be semantic consistent with exemplars.
Our method does not bring any extra inference cost and outperforms state-of-the-art methods on two benchmarks.
arXiv Detail & Related papers (2022-04-19T15:15:18Z) - Dominant Set-based Active Learning for Text Classification and its
Application to Online Social Media [0.0]
We present a novel pool-based active learning method for the training of large unlabeled corpus with minimum annotation cost.
Our proposed method does not have any parameters to be tuned, making it dataset-independent.
Our method achieves a higher performance in comparison to the state-of-the-art active learning strategies.
arXiv Detail & Related papers (2022-01-28T19:19:03Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z) - Contextual Diversity for Active Learning [9.546771465714876]
Large datasets restrict the use of deep convolutional neural networks (CNNs) for many practical applications.
We introduce the notion of contextual diversity that captures the confusion associated with spatially co-occurring classes.
Our studies show clear advantages of using contextual diversity for active learning.
arXiv Detail & Related papers (2020-08-13T07:04:15Z) - Deep Active Learning by Model Interpretability [7.3461534048332275]
In this paper, we introduce the linearly separable regions of samples to the problem of active learning.
We propose a novel Deep Active learning approach by Model Interpretability (DAMI)
To keep the maximal representativeness of the entire unlabeled data, DAMI tries to select and label samples on different linearly separable regions.
arXiv Detail & Related papers (2020-07-23T16:06:27Z) - Learning Interpretable and Discrete Representations with Adversarial
Training for Unsupervised Text Classification [87.28408260725138]
TIGAN learns to encode texts into two disentangled representations, including a discrete code and a continuous noise.
The extracted topical words for representing latent topics show that TIGAN learns coherent and highly interpretable topics.
arXiv Detail & Related papers (2020-04-28T02:53:59Z) - COLAM: Co-Learning of Deep Neural Networks and Soft Labels via
Alternating Minimization [60.07531696857743]
Co-Learns DNNs and soft labels through Alternating Minimization of two objectives.
We propose COLAM framework that Co-Learns DNNs and soft labels through Alternating Minimization of two objectives.
arXiv Detail & Related papers (2020-04-26T17:50:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.