Unified Risk Analysis for Weakly Supervised Learning
- URL: http://arxiv.org/abs/2309.08216v1
- Date: Fri, 15 Sep 2023 07:30:15 GMT
- Title: Unified Risk Analysis for Weakly Supervised Learning
- Authors: Chao-Kai Chiang, Masashi Sugiyama
- Abstract summary: We introduce a framework providing a comprehensive understanding and a unified methodology for weakly supervised learning.
The formulation component of the framework, leveraging a contamination perspective, provides a unified interpretation of how weak supervision is formed.
The analysis component of the framework, viewed as a decontamination process, provides a systematic method of conducting risk rewrite.
- Score: 65.75775694815172
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Among the flourishing research of weakly supervised learning (WSL), we
recognize the lack of a unified interpretation of the mechanism behind the
weakly supervised scenarios, let alone a systematic treatment of the risk
rewrite problem, a crucial step in the empirical risk minimization approach. In
this paper, we introduce a framework providing a comprehensive understanding
and a unified methodology for WSL. The formulation component of the framework,
leveraging a contamination perspective, provides a unified interpretation of
how weak supervision is formed and subsumes fifteen existing WSL settings. The
induced reduction graphs offer comprehensive connections over WSLs. The
analysis component of the framework, viewed as a decontamination process,
provides a systematic method of conducting risk rewrite. In addition to the
conventional inverse matrix approach, we devise a novel strategy called
marginal chain aiming to decontaminate distributions. We justify the
feasibility of the proposed framework by recovering existing rewrites reported
in the literature.
Related papers
- Eliciting Critical Reasoning in Retrieval-Augmented Language Models via Contrastive Explanations [4.697267141773321]
Retrieval-augmented generation (RAG) has emerged as a critical mechanism in contemporary NLP to support Large Language Models (LLMs) in systematically accessing richer factual context.
Recent studies have shown that LLMs still struggle to critically analyse RAG-based in-context information, a limitation that may lead to incorrect inferences and hallucinations.
In this paper, we investigate how to elicit critical reasoning in RAG via contrastive explanations.
arXiv Detail & Related papers (2024-10-30T10:11:53Z) - Hidden in Plain Text: Emergence & Mitigation of Steganographic Collusion in LLMs [0.600808022072121]
Collusion to the disadvantage of others has been identified as a central form of undesirable agent cooperation.
The use of information hiding (steganography) in agent communications could render collusion practically undetectable.
This underscores the need for evaluation frameworks to monitor and mitigate steganographic collusion capabilities.
arXiv Detail & Related papers (2024-10-02T16:18:33Z) - A Distributional Analogue to the Successor Representation [54.99439648059807]
This paper contributes a new approach for distributional reinforcement learning.
It elucidates a clean separation of transition structure and reward in the learning process.
As an illustration, we show that it enables zero-shot risk-sensitive policy evaluation.
arXiv Detail & Related papers (2024-02-13T15:35:24Z) - Data Poisoning for In-context Learning [49.77204165250528]
In-context learning (ICL) has been recognized for its innovative ability to adapt to new tasks.
This paper delves into the critical issue of ICL's susceptibility to data poisoning attacks.
We introduce ICLPoison, a specialized attacking framework conceived to exploit the learning mechanisms of ICL.
arXiv Detail & Related papers (2024-02-03T14:20:20Z) - Sparsity-Guided Holistic Explanation for LLMs with Interpretable
Inference-Time Intervention [53.896974148579346]
Large Language Models (LLMs) have achieved unprecedented breakthroughs in various natural language processing domains.
The enigmatic black-box'' nature of LLMs remains a significant challenge for interpretability, hampering transparent and accountable applications.
We propose a novel methodology anchored in sparsity-guided techniques, aiming to provide a holistic interpretation of LLMs.
arXiv Detail & Related papers (2023-12-22T19:55:58Z) - An Investigation of Representation and Allocation Harms in Contrastive
Learning [55.42336321517228]
We demonstrate that contrastive learning (CL) tends to collapse representations of minority groups with certain majority groups.
We refer to this phenomenon as representation harm and demonstrate it on image and text datasets using the corresponding popular CL methods.
We provide a theoretical explanation for representation harm using a neural block model that leads to a representational collapse in a contrastive learning setting.
arXiv Detail & Related papers (2023-10-02T19:25:37Z) - It begins with a boundary: A geometric view on probabilistically robust learning [6.877576704011329]
We take a fresh and geometric view on one such method -- Probabilistically Robust Learning (PRL)
We prove existence of solutions to the original and modified problems using novel relaxation methods.
We also clarify, through a suitable $Gamma$-convergence analysis, the way in which the original and modified PRL models interpolate between risk minimization and adversarial training.
arXiv Detail & Related papers (2023-05-30T06:24:30Z) - Revisiting Deep Semi-supervised Learning: An Empirical Distribution
Alignment Framework and Its Generalization Bound [97.93945601881407]
We propose a new deep semi-supervised learning framework called Semi-supervised Learning by Empirical Distribution Alignment (SLEDA)
We show the generalization error of semi-supervised learning can be effectively bounded by minimizing the training error on labeled data.
Building upon our new framework and the theoretical bound, we develop a simple and effective deep semi-supervised learning method called Augmented Distribution Alignment Network (ADA-Net)
arXiv Detail & Related papers (2022-03-13T11:59:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.