Does Head Label Help for Long-Tailed Multi-Label Text Classification
- URL: http://arxiv.org/abs/2101.09704v1
- Date: Sun, 24 Jan 2021 12:31:39 GMT
- Title: Does Head Label Help for Long-Tailed Multi-Label Text Classification
- Authors: Lin Xiao, Xiangliang Zhang, Liping Jing, Chi Huang, Mingyang Song
- Abstract summary: In real applications, the distribution of label frequency often exhibits a long tail, i.e., a few labels are associated with a large number of documents.
We propose a Head-to-Tail Network (HTTN) to transfer the meta-knowledge from the data-rich head labels to data-poor tail labels.
- Score: 45.762555329467446
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-label text classification (MLTC) aims to annotate documents with the
most relevant labels from a number of candidate labels. In real applications,
the distribution of label frequency often exhibits a long tail, i.e., a few
labels are associated with a large number of documents (a.k.a. head labels),
while a large fraction of labels are associated with a small number of
documents (a.k.a. tail labels). To address the challenge of insufficient
training data on tail label classification, we propose a Head-to-Tail Network
(HTTN) to transfer the meta-knowledge from the data-rich head labels to
data-poor tail labels. The meta-knowledge is the mapping from few-shot network
parameters to many-shot network parameters, which aims to promote the
generalizability of tail classifiers. Extensive experimental results on three
benchmark datasets demonstrate that HTTN consistently outperforms the
state-of-the-art methods. The code and hyper-parameter settings are released
for reproducibility
Related papers
- Weakly-Supervised Scientific Document Classification via
Retrieval-Augmented Multi-Stage Training [24.2734548438594]
We propose a weakly-supervised approach for scientific document classification using label names only.
In scientific domains, label names often include domain-specific concepts that may not appear in the document corpus.
We show that WANDER outperforms the best baseline by 11.9% on average.
arXiv Detail & Related papers (2023-06-12T15:50:13Z) - Exploring Structured Semantic Prior for Multi Label Recognition with
Incomplete Labels [60.675714333081466]
Multi-label recognition (MLR) with incomplete labels is very challenging.
Recent works strive to explore the image-to-label correspondence in the vision-language model, ie, CLIP, to compensate for insufficient annotations.
We advocate remedying the deficiency of label supervision for the MLR with incomplete labels by deriving a structured semantic prior.
arXiv Detail & Related papers (2023-03-23T12:39:20Z) - Review of Extreme Multilabel Classification [1.888738346075831]
Extreme multilabel classification or XML, is an active area of interest in machine learning.
The community has come up with a useful set of metrics to identify correctly the prediction for head or tail labels.
arXiv Detail & Related papers (2023-02-12T18:29:20Z) - Pairwise Instance Relation Augmentation for Long-tailed Multi-label Text
Classification [38.66674700075432]
We propose a Pairwise Instance Relation Augmentation Network (PIRAN) to augment tailed-label documents for balancing tail labels and head labels.
PIRAN consistently outperforms the SOTA methods, and dramatically improves the performance of tail labels.
arXiv Detail & Related papers (2022-11-19T12:45:54Z) - Structured Semantic Transfer for Multi-Label Recognition with Partial
Labels [85.6967666661044]
We propose a structured semantic transfer (SST) framework that enables training multi-label recognition models with partial labels.
The framework consists of two complementary transfer modules that explore within-image and cross-image semantic correlations.
Experiments on the Microsoft COCO, Visual Genome and Pascal VOC datasets show that the proposed SST framework obtains superior performance over current state-of-the-art algorithms.
arXiv Detail & Related papers (2021-12-21T02:15:01Z) - MATCH: Metadata-Aware Text Classification in A Large Hierarchy [60.59183151617578]
MATCH is an end-to-end framework that leverages both metadata and hierarchy information.
We propose different ways to regularize the parameters and output probability of each child label by its parents.
Experiments on two massive text datasets with large-scale label hierarchies demonstrate the effectiveness of MATCH.
arXiv Detail & Related papers (2021-02-15T05:23:08Z) - A Study on the Autoregressive and non-Autoregressive Multi-label
Learning [77.11075863067131]
We propose a self-attention based variational encoder-model to extract the label-label and label-feature dependencies jointly.
Our model can therefore be used to predict all labels in parallel while still including both label-label and label-feature dependencies.
arXiv Detail & Related papers (2020-12-03T05:41:44Z) - Label-Wise Document Pre-Training for Multi-Label Text Classification [14.439051753832032]
This paper develops Label-Wise Pre-Training (LW-PT) method to get a document representation with label-aware information.
The basic idea is that, a multi-label document can be represented as a combination of multiple label-wise representations, and that, correlated labels always cooccur in the same or similar documents.
arXiv Detail & Related papers (2020-08-15T10:34:27Z) - Few-shot Slot Tagging with Collapsed Dependency Transfer and
Label-enhanced Task-adaptive Projection Network [61.94394163309688]
We propose a Label-enhanced Task-Adaptive Projection Network (L-TapNet) based on the state-of-the-art few-shot classification model -- TapNet.
Experimental results show that our model significantly outperforms the strongest few-shot learning baseline by 14.64 F1 scores in the one-shot setting.
arXiv Detail & Related papers (2020-06-10T07:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.