Beyond Predictive Uncertainty: Reliable Representation Learning with Structural Constraints
- URL: http://arxiv.org/abs/2601.16174v1
- Date: Thu, 22 Jan 2026 18:19:52 GMT
- Title: Beyond Predictive Uncertainty: Reliable Representation Learning with Structural Constraints
- Authors: Yiyao Yang,
- Abstract summary: We argue that reliability should be regarded as a first-class property of learned representations themselves.<n>We propose a principled framework for reliable representation learning that explicitly models representation-level uncertainty.<n>Our approach introduces uncertainty-aware regularization directly in the representation space, encouraging representations that are not only predictive but also stable, well-calibrated, and robust to noise and structural perturbations.
- Score: 0.3948325938742681
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Uncertainty estimation in machine learning has traditionally focused on the prediction stage, aiming to quantify confidence in model outputs while treating learned representations as deterministic and reliable by default. In this work, we challenge this implicit assumption and argue that reliability should be regarded as a first-class property of learned representations themselves. We propose a principled framework for reliable representation learning that explicitly models representation-level uncertainty and leverages structural constraints as inductive biases to regularize the space of feasible representations. Our approach introduces uncertainty-aware regularization directly in the representation space, encouraging representations that are not only predictive but also stable, well-calibrated, and robust to noise and structural perturbations. Structural constraints, such as sparsity, relational structure, or feature-group dependencies, are incorporated to define meaningful geometry and reduce spurious variability in learned representations, without assuming fully correct or noise-free structure. Importantly, the proposed framework is independent of specific model architectures and can be integrated with a wide range of representation learning methods.
Related papers
- On the Identifiability of Steering Vectors in Large Language Models [0.0]
Activation steering methods are widely used to control large language model behavior.<n>This interpretation implicitly assumes steering directions are identifiable and uniquely recoverable from input-output behavior.<n>We prove that steering vectors are fundamentally non-identifiable due to large equivalence classes of behaviorally indistinguishable interventions.
arXiv Detail & Related papers (2026-02-06T15:53:50Z) - CURVE: Learning Causality-Inspired Invariant Representations for Robust Scene Understanding via Uncertainty-Guided Regularization [30.613712415224473]
CURVE is a framework that integrates variational uncertainty modeling with uncertainty-guided structural regularization to suppress high-variance relations.<n>Specifically, we apply prototype-conditioned debiasing to disentangle invariant interaction dynamics from environment-dependent variations, promoting a sparse and domain-stable topology.<n> Empirically, we evaluate CURVE in zero-shot transfer and low-data sim-to-real adaptation, verifying its ability to learn domain-stable sparse topologies and provide reliable uncertainty estimates to support risk prediction under distribution shifts.
arXiv Detail & Related papers (2026-01-28T08:15:56Z) - Coherence Mechanisms for Provable Self-Improvement [38.3455527898461]
We propose a principled framework for self-improvement based on the concept of emphcoherence<n>We formalize this concept using projection-based mechanisms that update a baseline model to be coherent while remaining as close as possible to its original behavior.<n>Our analysis is comprehensive, covering both emphdirect and emphtwo-step projection methods, and robustly extends these guarantees to non-realizable settings.
arXiv Detail & Related papers (2025-11-11T16:45:14Z) - Nonparametric Factor Analysis and Beyond [14.232694150264628]
We propose a general framework for identifying latent variables in the non-negligible settings.<n>We show that the generative model is identifiable up to certain submanifold indeterminacies even in the presence of non-negligible noise.<n>We have also developed corresponding estimation methods and validated them in various synthetic and real-world settings.
arXiv Detail & Related papers (2025-03-21T05:45:03Z) - Quantification of model error for inverse problems in the Weak Neural Variational Inference framework [0.0]
We present a novel extension to the Weak Neural Variational Inference (WNVI) framework for probabilistic property estimation.<n>Our framework explicitly quantifies model errors in PDE-based inverse problems.<n>Our findings suggest that the proposed framework enhances the accuracy and reliability of material property estimation.
arXiv Detail & Related papers (2025-02-11T09:52:06Z) - Know Where You're Uncertain When Planning with Multimodal Foundation Models: A Formal Framework [54.40508478482667]
We present a comprehensive framework to disentangle, quantify, and mitigate uncertainty in perception and plan generation.<n>We propose methods tailored to the unique properties of perception and decision-making.<n>We show that our uncertainty disentanglement framework reduces variability by up to 40% and enhances task success rates by 5% compared to baselines.
arXiv Detail & Related papers (2024-11-03T17:32:00Z) - The Foundations of Tokenization: Statistical and Computational Concerns [51.370165245628975]
Tokenization is a critical step in the NLP pipeline.<n>Despite its recognized importance as a standard representation method in NLP, the theoretical underpinnings of tokenization are not yet fully understood.<n>The present paper contributes to addressing this theoretical gap by proposing a unified formal framework for representing and analyzing tokenizer models.
arXiv Detail & Related papers (2024-07-16T11:12:28Z) - Quantification of Predictive Uncertainty via Inference-Time Sampling [57.749601811982096]
We propose a post-hoc sampling strategy for estimating predictive uncertainty accounting for data ambiguity.
The method can generate different plausible outputs for a given input and does not assume parametric forms of predictive distributions.
arXiv Detail & Related papers (2023-08-03T12:43:21Z) - Model-free generalized fiducial inference [0.0]
Conformal prediction (CP) was developed to provide finite-sample probabilistic prediction guarantees.<n>CP algorithms are a relatively general-purpose approach to uncertainty quantification, with finite-sample guarantees, they lack versatility.<n>In this paper, tools are offered from imprecise probability theory to build a formal connection between CP and generalized fiducial (GF) inference.
arXiv Detail & Related papers (2023-07-24T01:58:48Z) - When Does Confidence-Based Cascade Deferral Suffice? [69.28314307469381]
Cascades are a classical strategy to enable inference cost to vary adaptively across samples.
A deferral rule determines whether to invoke the next classifier in the sequence, or to terminate prediction.
Despite being oblivious to the structure of the cascade, confidence-based deferral often works remarkably well in practice.
arXiv Detail & Related papers (2023-07-06T04:13:57Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Quantifying Representation Reliability in Self-Supervised Learning Models [12.485580780944083]
Self-supervised learning models extract general-purpose representations from data.
We introduce a formal definition of representation reliability.
We propose an ensemble-based method for estimating the representation reliability without knowing the downstream tasks a priori.
arXiv Detail & Related papers (2023-05-31T21:57:33Z) - Trust but Verify: Assigning Prediction Credibility by Counterfactual
Constrained Learning [123.3472310767721]
Prediction credibility measures are fundamental in statistics and machine learning.
These measures should account for the wide variety of models used in practice.
The framework developed in this work expresses the credibility as a risk-fit trade-off.
arXiv Detail & Related papers (2020-11-24T19:52:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.