A Survey of Label-noise Representation Learning: Past, Present and
Future
- URL: http://arxiv.org/abs/2011.04406v2
- Date: Sat, 20 Feb 2021 07:28:12 GMT
- Title: A Survey of Label-noise Representation Learning: Past, Present and
Future
- Authors: Bo Han, Quanming Yao, Tongliang Liu, Gang Niu, Ivor W. Tsang, James T.
Kwok and Masashi Sugiyama
- Abstract summary: Label-Noise Representation Learning (LNRL) methods can robustly train deep models with noisy labels.
LNRL methods can be classified into three directions: instance-dependent LNRL, adversarial LNRL, and new datasets.
We envision potential directions beyond LNRL, such as learning with feature-noise, preference-noise, domain-noise, similarity-noise, graph-noise and demonstration-noise.
- Score: 172.28865582415628
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Classical machine learning implicitly assumes that labels of the training
data are sampled from a clean distribution, which can be too restrictive for
real-world scenarios. However, statistical-learning-based methods may not train
deep learning models robustly with these noisy labels. Therefore, it is urgent
to design Label-Noise Representation Learning (LNRL) methods for robustly
training deep models with noisy labels. To fully understand LNRL, we conduct a
survey study. We first clarify a formal definition for LNRL from the
perspective of machine learning. Then, via the lens of learning theory and
empirical study, we figure out why noisy labels affect deep models'
performance. Based on the theoretical guidance, we categorize different LNRL
methods into three directions. Under this unified taxonomy, we provide a
thorough discussion of the pros and cons of different categories. More
importantly, we summarize the essential components of robust LNRL, which can
spark new directions. Lastly, we propose possible research directions within
LNRL, such as new datasets, instance-dependent LNRL, and adversarial LNRL. We
also envision potential directions beyond LNRL, such as learning with
feature-noise, preference-noise, domain-noise, similarity-noise, graph-noise
and demonstration-noise.
Related papers
- LUNAR: LLM Unlearning via Neural Activation Redirection [20.60687563657169]
Large Language Models (LLMs) benefit from training on ever larger amounts of textual data, but they increasingly incur the risk of leaking private information.
We propose LUNAR, a novel unlearning methodology grounded in the Linear Representation Hypothesis.
We show that LUNAR achieves state-of-the-art unlearning performance while significantly enhancing the controllability of the unlearned model during inference.
arXiv Detail & Related papers (2025-02-11T03:23:22Z) - On the Nonlinearity of Layer Normalization [5.0464797863553414]
We investigate the representation capacity of a network with layerwise composition of linear and LN transformations, referred to as LN-Net.
We show that, given $m$ samples with any label assignment, an LN-Net with only 3 neurons in each layer and $O(m)$ LN layers can correctly classify them.
arXiv Detail & Related papers (2024-06-03T12:11:34Z) - How Can LLM Guide RL? A Value-Based Approach [68.55316627400683]
Reinforcement learning (RL) has become the de facto standard practice for sequential decision-making problems by improving future acting policies with feedback.
Recent developments in large language models (LLMs) have showcased impressive capabilities in language understanding and generation, yet they fall short in exploration and self-improvement capabilities.
We develop an algorithm named LINVIT that incorporates LLM guidance as a regularization factor in value-based RL, leading to significant reductions in the amount of data needed for learning.
arXiv Detail & Related papers (2024-02-25T20:07:13Z) - The RL/LLM Taxonomy Tree: Reviewing Synergies Between Reinforcement
Learning and Large Language Models [2.5721733711031978]
We review research studies that combine Reinforcement Learning (RL) and Large Language Models (LLMs)
We propose a novel taxonomy of three main classes based on the way that the two model types interact with each other.
arXiv Detail & Related papers (2024-02-02T20:01:15Z) - Disentangled Representation Learning [46.51815065323667]
Disentangled Representation Learning (DRL) aims to learn a model capable of identifying and disentangling the underlying factors hidden in the observable data in representation form.
We comprehensively investigate DRL from various aspects including motivations, definitions, methodologies, evaluations, applications, and model designs.
arXiv Detail & Related papers (2022-11-21T18:14:38Z) - Semi-Supervised Offline Reinforcement Learning with Action-Free
Trajectories [37.14064734165109]
Natural agents can learn from multiple data sources that differ in size, quality, and types of measurements.
We study this in the context of offline reinforcement learning (RL) by introducing a new, practically motivated semi-supervised setting.
arXiv Detail & Related papers (2022-10-12T18:22:23Z) - RvS: What is Essential for Offline RL via Supervised Learning? [77.91045677562802]
Recent work has shown that supervised learning alone, without temporal difference (TD) learning, can be remarkably effective for offline RL.
In every environment suite we consider simply maximizing likelihood with two-layer feedforward is competitive.
They also probe the limits of existing RvS methods, which are comparatively weak on random data.
arXiv Detail & Related papers (2021-12-20T18:55:16Z) - Neural Unsupervised Semantic Role Labeling [48.69930912510414]
We present the first neural unsupervised model for semantic role labeling.
We decompose the task as two argument related subtasks, identification and clustering.
Experiments on CoNLL-2009 English dataset demonstrate that our model outperforms previous state-of-the-art baseline.
arXiv Detail & Related papers (2021-04-19T04:50:16Z) - Syntax Role for Neural Semantic Role Labeling [77.5166510071142]
Semantic role labeling (SRL) is dedicated to recognizing the semantic predicate-argument structure of a sentence.
Previous studies in terms of traditional models have shown syntactic information can make remarkable contributions to SRL performance.
Recent neural SRL studies show that syntax information becomes much less important for neural semantic role labeling.
arXiv Detail & Related papers (2020-09-12T07:01:12Z) - Explainability in Deep Reinforcement Learning [68.8204255655161]
We review recent works in the direction to attain Explainable Reinforcement Learning (XRL)
In critical situations where it is essential to justify and explain the agent's behaviour, better explainability and interpretability of RL models could help gain scientific insight on the inner workings of what is still considered a black box.
arXiv Detail & Related papers (2020-08-15T10:11:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.