Two out of Three (ToT): using self-consistency to make robust predictions
- URL: http://arxiv.org/abs/2505.12642v1
- Date: Mon, 19 May 2025 02:50:19 GMT
- Title: Two out of Three (ToT): using self-consistency to make robust predictions
- Authors: Jung Hoon Lee, Sujith Vijayan,
- Abstract summary: We develop an algorithm that allows deep learning models to abstain from answering when they are uncertain.<n>Our algorithm, named Two out of Three (ToT)', is inspired by the sensitivity of the human brain to conflicting information.
- Score: 1.7314342339585087
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning (DL) can automatically construct intelligent agents, deep neural networks (alternatively, DL models), that can outperform humans in certain tasks. However, the operating principles of DL remain poorly understood, making its decisions incomprehensible. As a result, it poses a great risk to deploy DL in high-stakes domains in which mistakes or errors may lead to critical consequences. Here, we aim to develop an algorithm that can help DL models make more robust decisions by allowing them to abstain from answering when they are uncertain. Our algorithm, named `Two out of Three (ToT)', is inspired by the sensitivity of the human brain to conflicting information. ToT creates two alternative predictions in addition to the original model prediction and uses the alternative predictions to decide whether it should provide an answer or not.
Related papers
- Towards a Cascaded LLM Framework for Cost-effective Human-AI Decision-Making [55.2480439325792]
We present a cascaded LLM decision framework that adaptively delegates tasks across multiple tiers of expertise.<n>First, a deferral policy determines whether to accept the base model's answer or regenerate it with the large model.<n>Second, an abstention policy decides whether the cascade model response is sufficiently certain or requires human intervention.
arXiv Detail & Related papers (2025-06-13T15:36:22Z) - CoThink: Token-Efficient Reasoning via Instruct Models Guiding Reasoning Models [56.40065909544213]
Large language models (LLMs) benefit from increased test-time compute, a phenomenon known as test-time scaling.<n>However, reasoning-optimized models often overthink even simple problems, producing excessively verbose outputs and leading to low token efficiency.<n>We identify two key causes of this verbosity: (1) reinforcement learning reduces the information density of forward reasoning, and (2) backward chain-of thought training encourages redundant and often unnecessary verification steps.
arXiv Detail & Related papers (2025-05-28T06:24:45Z) - Modeling Boundedly Rational Agents with Latent Inference Budgets [56.24971011281947]
We introduce a latent inference budget model (L-IBM) that models agents' computational constraints explicitly.
L-IBMs make it possible to learn agent models using data from diverse populations of suboptimal actors.
We show that L-IBMs match or outperform Boltzmann models of decision-making under uncertainty.
arXiv Detail & Related papers (2023-12-07T03:55:51Z) - Interpretable Imitation Learning with Dynamic Causal Relations [65.18456572421702]
We propose to expose captured knowledge in the form of a directed acyclic causal graph.
We also design this causal discovery process to be state-dependent, enabling it to model the dynamics in latent causal graphs.
The proposed framework is composed of three parts: a dynamic causal discovery module, a causality encoding module, and a prediction module, and is trained in an end-to-end manner.
arXiv Detail & Related papers (2023-09-30T20:59:42Z) - Interpretable Self-Aware Neural Networks for Robust Trajectory
Prediction [50.79827516897913]
We introduce an interpretable paradigm for trajectory prediction that distributes the uncertainty among semantic concepts.
We validate our approach on real-world autonomous driving data, demonstrating superior performance over state-of-the-art baselines.
arXiv Detail & Related papers (2022-11-16T06:28:20Z) - Autoregressive Uncertainty Modeling for 3D Bounding Box Prediction [63.3021778885906]
3D bounding boxes are a widespread intermediate representation in many computer vision applications.
We propose methods for leveraging our autoregressive model to make high confidence predictions and meaningful uncertainty measures.
We release a simulated dataset, COB-3D, which highlights new types of ambiguity that arise in real-world robotics applications.
arXiv Detail & Related papers (2022-10-13T23:57:40Z) - Generalization ability and Vulnerabilities to adversarial perturbations: Two sides of the same coin [1.7314342339585087]
We use the self-organizing map (SOM) to analyze internal codes associated with deep learning models' decision-making.
Our analyses suggest that shallow layers close to the input layer map onto homogeneous codes and that deep layers close to the output layer transform these homogeneous codes in shallow layers to diverse codes.
arXiv Detail & Related papers (2022-05-22T23:14:27Z) - Explainable Artificial Intelligence for Bayesian Neural Networks:
Towards trustworthy predictions of ocean dynamics [0.0]
The trustworthiness of neural networks is often challenged because they lack the ability to express uncertainty and explain their skill.
This can be problematic given the increasing use of neural networks in high stakes decision-making such as in climate change applications.
We address both issues by successfully implementing a Bayesian Neural Network (BNN), where parameters are distributions rather than deterministic, and applying novel implementations of explainable AI (XAI) techniques.
arXiv Detail & Related papers (2022-04-30T08:35:57Z) - Deep Learning Reproducibility and Explainable AI (XAI) [9.13755431537592]
The nondeterminism of Deep Learning (DL) training algorithms and its influence on the explainability of neural network (NN) models are investigated.
To discuss the issue, two convolutional neural networks (CNN) have been trained and their results compared.
arXiv Detail & Related papers (2022-02-23T12:06:20Z) - Towards Training Reproducible Deep Learning Models [26.547756923322126]
Deep Learning (DL) models are challenging to be reproduced due to issues like randomness in the software and non-determinism in the hardware.
This paper proposes a systematic approach to training reproducible DL models.
Case study results show our approach can successfully reproduce six open source and one commercial DL models.
arXiv Detail & Related papers (2022-02-04T18:14:39Z) - Prediction Surface Uncertainty Quantification in Object Detection Models
for Autonomous Driving [5.784950275336468]
Object detection in autonomous cars is commonly based on camera images and Lidar inputs, which are often used to train prediction models.
We propose a novel method called PURE (Prediction sURface uncErtainty) for measuring prediction uncertainty of such regression models.
arXiv Detail & Related papers (2021-07-11T08:31:15Z) - NBDT: Neural-Backed Decision Trees [26.2115887956431]
We improve accuracy and interpretability using Neural-Backed Decision Trees (NBDTs)
NBDTs replace a neural network's final linear layer with a differentiable sequence of decisions and a surrogate loss.
Our surrogate loss improves the original model's accuracy by up to 2%.
arXiv Detail & Related papers (2020-04-01T04:04:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.