Learnability with Indirect Supervision Signals
- URL: http://arxiv.org/abs/2006.08791v2
- Date: Wed, 11 Nov 2020 10:12:53 GMT
- Title: Learnability with Indirect Supervision Signals
- Authors: Kaifu Wang, Qiang Ning, Dan Roth
- Abstract summary: Learning from indirect supervision signals is important in real-world AI applications when, often, gold labels are missing or too costly.
We develop a unified theoretical framework for multi-class classification when the supervision is provided by a variable that contains nonzero mutual information with the gold label.
Our framework relaxes assumptions made in the literature, and supports learning with unknown, non-invertible and instance-dependent transitions.
- Score: 74.39088325025378
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning from indirect supervision signals is important in real-world AI
applications when, often, gold labels are missing or too costly. In this paper,
we develop a unified theoretical framework for multi-class classification when
the supervision is provided by a variable that contains nonzero mutual
information with the gold label. The nature of this problem is determined by
(i) the transition probability from the gold labels to the indirect supervision
variables and (ii) the learner's prior knowledge about the transition. Our
framework relaxes assumptions made in the literature, and supports learning
with unknown, non-invertible and instance-dependent transitions. Our theory
introduces a novel concept called \emph{separation}, which characterizes the
learnability and generalization bounds. We also demonstrate the application of
our framework via concrete novel results in a variety of learning scenarios
such as learning with superset annotations and joint supervision signals.
Related papers
- On Learning Latent Models with Multi-Instance Weak Supervision [57.18649648182171]
We consider a weakly supervised learning scenario where the supervision signal is generated by a transition function $sigma$ labels associated with multiple input instances.
Our problem is met in different fields, including latent structural learning and neuro-symbolic integration.
arXiv Detail & Related papers (2023-06-23T22:05:08Z) - PatchMix Augmentation to Identify Causal Features in Few-shot Learning [55.64873998196191]
Few-shot learning aims to transfer knowledge learned from base with sufficient categories labelled data to novel categories with scarce known information.
We propose a novel data augmentation strategy dubbed as PatchMix that can break this spurious dependency.
We show that such an augmentation mechanism, different from existing ones, is able to identify the causal features.
arXiv Detail & Related papers (2022-11-29T08:41:29Z) - Label Propagation with Weak Supervision [47.52032178837098]
We introduce a novel analysis of the classical label propagation algorithm (LPA) (Zhu & Ghahramani, 2002)
We provide an error bound that exploits both the local geometric properties of the underlying graph and the quality of the prior information.
We demonstrate the ability of our approach on multiple benchmark weakly supervised classification tasks, showing improvements upon existing semi-supervised and weakly supervised methods.
arXiv Detail & Related papers (2022-10-07T14:53:02Z) - Creating Training Sets via Weak Indirect Supervision [66.77795318313372]
Weak Supervision (WS) frameworks synthesize training labels from multiple potentially noisy supervision sources.
We formulate Weak Indirect Supervision (WIS), a new research problem for automatically synthesizing training labels.
We develop a probabilistic modeling approach, PLRM, which uses user-provided label relations to model and leverage indirect supervision sources.
arXiv Detail & Related papers (2021-10-07T14:09:35Z) - A Review of Open-World Learning and Steps Toward Open-World Learning
Without Labels [11.380522815465984]
In open-world learning, an agent starts with a set of known classes, detects, and manages things that it does not know, and learns them over time from a non-stationary stream of data.
This paper formalizes various open-world learning problems including open-world learning without labels.
arXiv Detail & Related papers (2020-11-25T17:41:03Z) - Foreseeing the Benefits of Incidental Supervision [83.08441990812636]
This paper studies whether we can, in a single framework, quantify the benefits of various types of incidental signals for a given target task without going through experiments.
We propose a unified PAC-Bayesian motivated informativeness measure, PABI, that characterizes the uncertainty reduction provided by incidental supervision signals.
arXiv Detail & Related papers (2020-06-09T20:59:42Z) - Structured Prediction with Partial Labelling through the Infimum Loss [85.4940853372503]
The goal of weak supervision is to enable models to learn using only forms of labelling which are cheaper to collect.
This is a type of incomplete annotation where, for each datapoint, supervision is cast as a set of labels containing the real one.
This paper provides a unified framework based on structured prediction and on the concept of infimum loss to deal with partial labelling.
arXiv Detail & Related papers (2020-03-02T13:59:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.