Label Noise-Resistant Mean Teaching for Weakly Supervised Fake News
Detection
- URL: http://arxiv.org/abs/2206.12260v1
- Date: Fri, 10 Jun 2022 16:01:58 GMT
- Title: Label Noise-Resistant Mean Teaching for Weakly Supervised Fake News
Detection
- Authors: Jingyi Xie, Jiawei Liu, Zheng-Jun Zha
- Abstract summary: We propose a novel label noise-resistant mean teaching approach (LNMT) for weakly supervised fake news detection.
LNMT leverages unlabeled news and feedback comments of users to enlarge the amount of training data.
LNMT establishes a mean teacher framework equipped with label propagation and label reliability estimation.
- Score: 93.6222609806278
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fake news spreads at an unprecedented speed, reaches global audiences and
poses huge risks to users and communities. Most existing fake news detection
algorithms focus on building supervised training models on a large amount of
manually labeled data, which is expensive to acquire or often unavailable. In
this work, we propose a novel label noise-resistant mean teaching approach
(LNMT) for weakly supervised fake news detection. LNMT leverages unlabeled news
and feedback comments of users to enlarge the amount of training data and
facilitates model training by generating refined labels as weak supervision.
Specifically, LNMT automatically assigns initial weak labels to unlabeled
samples based on semantic correlation and emotional association between news
content and the comments. Moreover, in order to suppress the noises in weak
labels, LNMT establishes a mean teacher framework equipped with label
propagation and label reliability estimation. The framework measures a weak
label similarity matrix between the teacher and student networks, and
propagates different valuable weak label information to refine the weak labels.
Meanwhile, it exploits the consistency between the output class likelihood
vectors of the two networks to evaluate the reliability of the weak labels and
incorporates the reliability into model optimization to alleviate the negative
effect of noisy weak labels. Extensive experiments show the superior
performance of LNMT.
Related papers
- BadLabel: A Robust Perspective on Evaluating and Enhancing Label-noise
Learning [113.8799653759137]
We introduce a novel label noise type called BadLabel, which can significantly degrade the performance of existing LNL algorithms by a large margin.
BadLabel is crafted based on the label-flipping attack against standard classification.
We propose a robust LNL method that perturbs the labels in an adversarial manner at each epoch to make the loss values of clean and noisy labels again distinguishable.
arXiv Detail & Related papers (2023-05-28T06:26:23Z) - Learning from Noisy Labels with Decoupled Meta Label Purifier [33.87292143223425]
Training deep neural networks with noisy labels is challenging since DNN can easily memorize inaccurate labels.
In this paper, we propose a novel multi-stage label purifier named DMLP.
DMLP decouples the label correction process into label-free representation learning and a simple meta label purifier.
arXiv Detail & Related papers (2023-02-14T03:39:30Z) - Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective [89.5370481649529]
We propose a label distribution perspective for PU learning in this paper.
Motivated by this, we propose to pursue the label distribution consistency between predicted and ground-truth label distributions.
Experiments on three benchmark datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2022-12-06T07:38:29Z) - Transductive CLIP with Class-Conditional Contrastive Learning [68.51078382124331]
We propose Transductive CLIP, a novel framework for learning a classification network with noisy labels from scratch.
A class-conditional contrastive learning mechanism is proposed to mitigate the reliance on pseudo labels.
ensemble labels is adopted as a pseudo label updating strategy to stabilize the training of deep neural networks with noisy labels.
arXiv Detail & Related papers (2022-06-13T14:04:57Z) - CLS: Cross Labeling Supervision for Semi-Supervised Learning [9.929229055862491]
Cross Labeling Supervision ( CLS) is a framework that generalizes the typical pseudo-labeling process.
CLS allows the creation of both pseudo and complementary labels to support both positive and negative learning.
arXiv Detail & Related papers (2022-02-17T08:09:40Z) - Data Consistency for Weakly Supervised Learning [15.365232702938677]
Training machine learning models involves using large amounts of human-annotated data.
We propose a novel weak supervision algorithm that processes noisy labels, i.e., weak signals.
We show that it significantly outperforms state-of-the-art weak supervision methods on both text and image classification tasks.
arXiv Detail & Related papers (2022-02-08T16:48:19Z) - Informative Pseudo-Labeling for Graph Neural Networks with Few Labels [12.83841767562179]
Graph Neural Networks (GNNs) have achieved state-of-the-art results for semi-supervised node classification on graphs.
The challenge of how to effectively learn GNNs with very few labels is still under-explored.
We propose a novel informative pseudo-labeling framework, called InfoGNN, to facilitate learning of GNNs with extremely few labels.
arXiv Detail & Related papers (2022-01-20T01:49:30Z) - Boosting Semi-Supervised Face Recognition with Noise Robustness [54.342992887966616]
This paper presents an effective solution to semi-supervised face recognition that is robust to the label noise aroused by the auto-labelling.
We develop a semi-supervised face recognition solution, named Noise Robust Learning-Labelling (NRoLL), which is based on the robust training ability empowered by GN.
arXiv Detail & Related papers (2021-05-10T14:43:11Z) - Denoising Multi-Source Weak Supervision for Neural Text Classification [9.099703420721701]
We study the problem of learning neural text classifiers without using any labeled data, but only easy-to-provide rules as multiple weak supervision sources.
This problem is challenging because rule-induced weak labels are often noisy and incomplete.
We design a label denoiser, which estimates the source reliability using a conditional soft attention mechanism and then reduces label noise by aggregating rule-annotated weak labels.
arXiv Detail & Related papers (2020-10-09T13:57:52Z) - Semi-supervised Relation Extraction via Incremental Meta Self-Training [56.633441255756075]
Semi-Supervised Relation Extraction methods aim to leverage unlabeled data in addition to learning from limited samples.
Existing self-training methods suffer from the gradual drift problem, where noisy pseudo labels on unlabeled data are incorporated during training.
We propose a method called MetaSRE, where a Relation Label Generation Network generates quality assessment on pseudo labels by (meta) learning from the successful and failed attempts on Relation Classification Network as an additional meta-objective.
arXiv Detail & Related papers (2020-10-06T03:54:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.