R-U-SURE? Uncertainty-Aware Code Suggestions By Maximizing Utility
Across Random User Intents
- URL: http://arxiv.org/abs/2303.00732v2
- Date: Fri, 28 Apr 2023 22:36:09 GMT
- Title: R-U-SURE? Uncertainty-Aware Code Suggestions By Maximizing Utility
Across Random User Intents
- Authors: Daniel D. Johnson, Daniel Tarlow, Christian Walder
- Abstract summary: Large language models show impressive results at predicting structured text such as code, but also commonly introduce errors and hallucinations in their output.
We propose Randomized Utility-driven Synthesis of Uncertain REgions (R-U-SURE)
R-U-SURE is an approach for building uncertainty-aware suggestions based on a decision-theoretic model of goal-conditioned utility.
- Score: 14.455036827804541
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large language models show impressive results at predicting structured text
such as code, but also commonly introduce errors and hallucinations in their
output. When used to assist software developers, these models may make mistakes
that users must go back and fix, or worse, introduce subtle bugs that users may
miss entirely. We propose Randomized Utility-driven Synthesis of Uncertain
REgions (R-U-SURE), an approach for building uncertainty-aware suggestions
based on a decision-theoretic model of goal-conditioned utility, using random
samples from a generative model as a proxy for the unobserved possible intents
of the end user. Our technique combines minimum-Bayes-risk decoding, dual
decomposition, and decision diagrams in order to efficiently produce structured
uncertainty summaries, given only sample access to an arbitrary generative
model of code and an optional AST parser. We demonstrate R-U-SURE on three
developer-assistance tasks, and show that it can be applied different user
interaction patterns without retraining the model and leads to more accurate
uncertainty estimates than token-probability baselines. We also release our
implementation as an open-source library at
https://github.com/google-research/r_u_sure.
Related papers
- Unsupervised Model Diagnosis [49.36194740479798]
This paper proposes Unsupervised Model Diagnosis (UMO) to produce semantic counterfactual explanations without any user guidance.
Our approach identifies and visualizes changes in semantics, and then matches these changes to attributes from wide-ranging text sources.
arXiv Detail & Related papers (2024-10-08T17:59:03Z) - A Probabilistic Perspective on Unlearning and Alignment for Large Language Models [48.96686419141881]
We introduce the first formal probabilistic evaluation framework in Large Language Models (LLMs)
We derive novel metrics with high-probability guarantees concerning the output distribution of a model.
Our metrics are application-independent and allow practitioners to make more reliable estimates about model capabilities before deployment.
arXiv Detail & Related papers (2024-10-04T15:44:23Z) - Rethinking Missing Data: Aleatoric Uncertainty-Aware Recommendation [59.500347564280204]
We propose a new Aleatoric Uncertainty-aware Recommendation (AUR) framework.
AUR consists of a new uncertainty estimator along with a normal recommender model.
As the chance of mislabeling reflects the potential of a pair, AUR makes recommendations according to the uncertainty.
arXiv Detail & Related papers (2022-09-22T04:32:51Z) - Uncertainty Estimation for Language Reward Models [5.33024001730262]
Language models can learn a range of capabilities from unsupervised training on text corpora.
It is often easier for humans to choose between options than to provide labeled data, and prior work has achieved state-of-the-art performance by training a reward model from such preference comparisons.
We seek to address these problems via uncertainty estimation, which can improve sample efficiency and robustness using active learning and risk-averse reinforcement learning.
arXiv Detail & Related papers (2022-03-14T20:13:21Z) - Probabilistic Modeling for Human Mesh Recovery [73.11532990173441]
This paper focuses on the problem of 3D human reconstruction from 2D evidence.
We recast the problem as learning a mapping from the input to a distribution of plausible 3D poses.
arXiv Detail & Related papers (2021-08-26T17:55:11Z) - A Gentle Introduction to Conformal Prediction and Distribution-Free
Uncertainty Quantification [1.90365714903665]
This hands-on introduction is aimed at a reader interested in the practical implementation of distribution-free UQ.
We will include many explanatory illustrations, examples, and code samples in Python, with PyTorch syntax.
arXiv Detail & Related papers (2021-07-15T17:59:50Z) - Multi-output Gaussian Processes for Uncertainty-aware Recommender
Systems [3.908842679355254]
We introduce an efficient strategy for model training and inference, resulting in a model that scales to very large and sparse datasets.
Our model also provides meaningful uncertainty estimates about quantifying that prediction.
arXiv Detail & Related papers (2021-06-08T10:01:14Z) - VAE Approximation Error: ELBO and Conditional Independence [78.72292013299868]
This paper analyzes VAE approximation errors caused by the combination of the ELBO objective with the choice of the encoder probability family.
We show that the ELBO subset can not be enlarged, and the respective error cannot be decreased, by only considering deeper encoder networks.
arXiv Detail & Related papers (2021-02-18T12:54:42Z) - Bayes DistNet -- A Robust Neural Network for Algorithm Runtime
Distribution Predictions [1.8275108630751844]
Randomized algorithms are used in many state-of-the-art solvers for constraint satisfaction problems (CSP) and Boolean satisfiability (SAT) problems.
Previous state-of-the-art methods directly try to predict a fixed parametric distribution that the input instance follows.
This new model achieves robust predictive performance in the low observation setting, as well as handling censored observations.
arXiv Detail & Related papers (2020-12-14T01:15:39Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.