Learning from Rules Generalizing Labeled Exemplars
- URL: http://arxiv.org/abs/2004.06025v2
- Date: Fri, 15 May 2020 15:56:59 GMT
- Title: Learning from Rules Generalizing Labeled Exemplars
- Authors: Abhijeet Awasthi, Sabyasachi Ghosh, Rasna Goyal, Sunita Sarawagi
- Abstract summary: In many applications labeled data is not readily available, and needs to be collected via pain-staking human supervision.
We propose a rule-exemplar method for collecting human supervision to combine the efficiency of rules with the quality of instance labels.
- Score: 21.359456842579945
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many applications labeled data is not readily available, and needs to be
collected via pain-staking human supervision. We propose a rule-exemplar method
for collecting human supervision to combine the efficiency of rules with the
quality of instance labels. The supervision is coupled such that it is both
natural for humans and synergistic for learning. We propose a training
algorithm that jointly denoises rules via latent coverage variables, and trains
the model through a soft implication loss over the coverage and label
variables. The denoised rules and trained model are used jointly for inference.
Empirical evaluation on five different tasks shows that (1) our algorithm is
more accurate than several existing methods of learning from a mix of clean and
noisy supervision, and (2) the coupled rule-exemplar supervision is effective
in denoising rules.
Related papers
- RuleAgent: Discovering Rules for Recommendation Denoising with Autonomous Language Agents [36.31706728494194]
RuleAgent mimics real-world data experts to autonomously discover rules for recommendation denoising.
LossEraser-an unlearning strategy streamlines training without compromising denoising performance.
arXiv Detail & Related papers (2025-03-30T09:19:03Z) - One-bit Supervision for Image Classification: Problem, Solution, and
Beyond [114.95815360508395]
This paper presents one-bit supervision, a novel setting of learning with fewer labels, for image classification.
We propose a multi-stage training paradigm and incorporate negative label suppression into an off-the-shelf semi-supervised learning algorithm.
In multiple benchmarks, the learning efficiency of the proposed approach surpasses that using full-bit, semi-supervised supervision.
arXiv Detail & Related papers (2023-11-26T07:39:00Z) - Enhancing Adversarial Robustness in Low-Label Regime via Adaptively
Weighted Regularization and Knowledge Distillation [1.675857332621569]
We investigate semi-supervised adversarial training where labeled data is scarce.
We develop a semi-supervised adversarial training algorithm that combines the proposed regularization term with knowledge distillation.
Our proposed algorithm achieves state-of-the-art performance with significant margins compared to existing algorithms.
arXiv Detail & Related papers (2023-08-08T05:48:38Z) - Hierarchical Semi-Supervised Contrastive Learning for
Contamination-Resistant Anomaly Detection [81.07346419422605]
Anomaly detection aims at identifying deviant samples from the normal data distribution.
Contrastive learning has provided a successful way to sample representation that enables effective discrimination on anomalies.
We propose a novel hierarchical semi-supervised contrastive learning framework, for contamination-resistant anomaly detection.
arXiv Detail & Related papers (2022-07-24T18:49:26Z) - Learning with Neighbor Consistency for Noisy Labels [69.83857578836769]
We present a method for learning from noisy labels that leverages similarities between training examples in feature space.
We evaluate our method on datasets evaluating both synthetic (CIFAR-10, CIFAR-100) and realistic (mini-WebVision, Clothing1M, mini-ImageNet-Red) noise.
arXiv Detail & Related papers (2022-02-04T15:46:27Z) - A Good Representation Detects Noisy Labels [9.4092903583089]
Label noise is pervasive in real-world datasets, which encodes wrong correlation patterns and impairs the generalization of deep neural networks (DNNs)
We propose a universally applicable and trainingfree solution to detect noisy labels.
Experiments with both synthetic and real-world label noise demonstrate our training-free solutions are significantly improving over most of the training-based datasets.
arXiv Detail & Related papers (2021-10-12T19:10:30Z) - Denoising Multi-Source Weak Supervision for Neural Text Classification [9.099703420721701]
We study the problem of learning neural text classifiers without using any labeled data, but only easy-to-provide rules as multiple weak supervision sources.
This problem is challenging because rule-induced weak labels are often noisy and incomplete.
We design a label denoiser, which estimates the source reliability using a conditional soft attention mechanism and then reduces label noise by aggregating rule-annotated weak labels.
arXiv Detail & Related papers (2020-10-09T13:57:52Z) - Noisy Concurrent Training for Efficient Learning under Label Noise [13.041607703862724]
Deep neural networks (DNNs) fail to learn effectively under label noise and have been shown to memorize random labels which affect their performance.
We consider learning in isolation, using one-hot encoded labels as the sole source of supervision, and a lack of regularization to discourage memorization as the major shortcomings of the standard training procedure.
We propose Noisy Concurrent Training (NCT) which leverages collaborative learning to use the consensus between two models as an additional source of supervision.
arXiv Detail & Related papers (2020-09-17T14:22:17Z) - Constrained Labeling for Weakly Supervised Learning [15.365232702938677]
We propose a simple data-free approach for combining weak supervision signals.
Our method is efficient and stable, converging after a few iterations of descent.
We show experimentally that our method outperforms other weak supervision methods on various text- and image-classification tasks.
arXiv Detail & Related papers (2020-09-15T21:30:53Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z) - Learning from Noisy Similar and Dissimilar Data [84.76686918337134]
We show how to learn a classifier from noisy S and D labeled data.
We also show important connections between learning from such pairwise supervision data and learning from ordinary class-labeled data.
arXiv Detail & Related papers (2020-02-03T19:59:16Z) - Hierarchical Variational Imitation Learning of Control Programs [131.7671843857375]
We propose a variational inference method for imitation learning of a control policy represented by parametrized hierarchical procedures (PHP)
Our method discovers the hierarchical structure in a dataset of observation-action traces of teacher demonstrations, by learning an approximate posterior distribution over the latent sequence of procedure calls and terminations.
We demonstrate a novel benefit of variational inference in the context of hierarchical imitation learning: in decomposing the policy into simpler procedures, inference can leverage acausal information that is unused by other methods.
arXiv Detail & Related papers (2019-12-29T08:57:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.