Interactive Concept Bottleneck Models
- URL: http://arxiv.org/abs/2212.07430v3
- Date: Thu, 27 Apr 2023 17:32:18 GMT
- Title: Interactive Concept Bottleneck Models
- Authors: Kushal Chauhan, Rishabh Tiwari, Jan Freyberg, Pradeep Shenoy,
Krishnamurthy Dvijotham
- Abstract summary: Concept bottleneck models (CBMs) are interpretable neural networks that first predict labels for human-interpretable concepts relevant to the prediction task.
We extend CBMs to interactive prediction settings where the model can query a human collaborator for the label to some concepts.
We develop an interaction policy that, at prediction time, chooses which concepts to request a label for so as to maximally improve the final prediction.
- Score: 14.240165842615674
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Concept bottleneck models (CBMs) are interpretable neural networks that first
predict labels for human-interpretable concepts relevant to the prediction
task, and then predict the final label based on the concept label predictions.
We extend CBMs to interactive prediction settings where the model can query a
human collaborator for the label to some concepts. We develop an interaction
policy that, at prediction time, chooses which concepts to request a label for
so as to maximally improve the final prediction. We demonstrate that a simple
policy combining concept prediction uncertainty and influence of the concept on
the final prediction achieves strong performance and outperforms static
approaches as well as active feature acquisition methods proposed in the
literature. We show that the interactive CBM can achieve accuracy gains of
5-10% with only 5 interactions over competitive baselines on the Caltech-UCSD
Birds, CheXpert and OAI datasets.
Related papers
- LLM Pretraining with Continuous Concepts [71.98047075145249]
Next token prediction has been the standard training objective used in large language model pretraining.
We propose Continuous Concept Mixing (CoCoMix), a novel pretraining framework that combines discrete next token prediction with continuous concepts.
arXiv Detail & Related papers (2025-02-12T16:00:11Z) - Survival Concept-Based Learning Models [2.024925013349319]
Two novel models are proposed to integrate concept-based learning with survival analysis.
SurvCBM is based on the architecture of the well-known concept bottleneck model.
SurvRCM uses concepts as regularization to enhance accuracy.
arXiv Detail & Related papers (2025-02-09T16:41:04Z) - Diverse Concept Proposals for Concept Bottleneck Models [23.395270888378594]
Concept bottleneck models are interpretable predictive models that are often used in domains where model trust is a key priority, such as healthcare.
Our proposed approach identifies a number of predictive concepts that explain the data.
By offering multiple alternative explanations, we allow the human expert to choose the one that best aligns with their expectation.
arXiv Detail & Related papers (2024-12-24T00:12:34Z) - Stochastic Concept Bottleneck Models [8.391254800873599]
Concept Bottleneck Models (CBMs) have emerged as a promising interpretable method whose final prediction is based on human-understandable concepts.
We propose Concept Bottleneck Models (SCBMs), a novel approach that models concept dependencies.
A single-concept intervention affects all correlated concepts, thereby improving intervention effectiveness.
arXiv Detail & Related papers (2024-06-27T15:38:37Z) - Certified Human Trajectory Prediction [66.1736456453465]
Tray prediction plays an essential role in autonomous vehicles.
We propose a certification approach tailored for the task of trajectory prediction.
We address the inherent challenges associated with trajectory prediction, including unbounded outputs, and mutli-modality.
arXiv Detail & Related papers (2024-03-20T17:41:35Z) - Can we Constrain Concept Bottleneck Models to Learn Semantically Meaningful Input Features? [0.6401548653313325]
Concept Bottleneck Models (CBMs) are regarded as inherently interpretable because they first predict a set of human-defined concepts.
Current literature suggests that concept predictions often rely on irrelevant input features.
In this paper, we demonstrate that CBMs can learn to map concepts to semantically meaningful input features.
arXiv Detail & Related papers (2024-02-01T10:18:43Z) - Conditional Goal-oriented Trajectory Prediction for Interacting Vehicles
with Vectorized Representation [10.190939530193502]
Conditional Goal-oriented Trajectory Prediction (CGTP) framework to jointly generate scene-compliant trajectories of two interacting agents.
Three main stages: context encoding, goal interactive prediction and trajectory interactive prediction.
New goal interactive loss is developed to better learn the joint probability distribution over goal candidates between two interacting agents.
arXiv Detail & Related papers (2022-10-19T15:14:46Z) - Predictive Inference with Feature Conformal Prediction [80.77443423828315]
We propose feature conformal prediction, which extends the scope of conformal prediction to semantic feature spaces.
From a theoretical perspective, we demonstrate that feature conformal prediction provably outperforms regular conformal prediction under mild assumptions.
Our approach could be combined with not only vanilla conformal prediction, but also other adaptive conformal prediction methods.
arXiv Detail & Related papers (2022-10-01T02:57:37Z) - Uncertainty estimation of pedestrian future trajectory using Bayesian
approximation [137.00426219455116]
Under dynamic traffic scenarios, planning based on deterministic predictions is not trustworthy.
The authors propose to quantify uncertainty during forecasting using approximation which deterministic approaches fail to capture.
The effect of dropout weights and long-term prediction on future state uncertainty has been studied.
arXiv Detail & Related papers (2022-05-04T04:23:38Z) - Online Multi-Agent Forecasting with Interpretable Collaborative Graph
Neural Network [65.11999700562869]
We propose a novel collaborative prediction unit (CoPU), which aggregates predictions from multiple collaborative predictors according to a collaborative graph.
Our methods outperform state-of-the-art works on the three tasks by 28.6%, 17.4% and 21.0% on average.
arXiv Detail & Related papers (2021-07-02T08:20:06Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.