Towards learning to explain with concept bottleneck models: mitigating
information leakage
- URL: http://arxiv.org/abs/2211.03656v1
- Date: Mon, 7 Nov 2022 16:10:36 GMT
- Title: Towards learning to explain with concept bottleneck models: mitigating
information leakage
- Authors: Joshua Lockhart, Nicolas Marchesotti, Daniele Magazzeni, Manuela
Veloso
- Abstract summary: Concept bottleneck models perform classification by first predicting which of a list of human provided concepts are true about a datapoint.
A downstream model uses these predicted concept labels to predict the target label.
The predicted concepts act as a rationale for the target prediction.
- Score: 19.52933192442871
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Concept bottleneck models perform classification by first predicting which of
a list of human provided concepts are true about a datapoint. Then a downstream
model uses these predicted concept labels to predict the target label. The
predicted concepts act as a rationale for the target prediction. Model trust
issues emerge in this paradigm when soft concept labels are used: it has
previously been observed that extra information about the data distribution
leaks into the concept predictions. In this work we show how Monte-Carlo
Dropout can be used to attain soft concept predictions that do not contain
leaked information.
Related papers
- MulCPred: Learning Multi-modal Concepts for Explainable Pedestrian Action Prediction [57.483718822429346]
MulCPred is proposed that explains its predictions based on multi-modal concepts represented by training samples.
MulCPred is evaluated on multiple datasets and tasks.
arXiv Detail & Related papers (2024-09-14T14:15:28Z) - Achieving Data Efficient Neural Networks with Hybrid Concept-based Models [0.0]
We introduce two novel model architectures that train using both class labels and additional information in the dataset referred to as concepts.
We show that the hybrid concept-based models outperform standard computer vision models with respect to accuracy, especially in sparse data settings.
We also introduce an algorithm for performing adversarial concept attacks, where an image is perturbed in a way that does not change a concept-based model's concept predictions, but changes the class prediction.
arXiv Detail & Related papers (2024-08-14T10:15:34Z) - Discover-then-Name: Task-Agnostic Concept Bottlenecks via Automated Concept Discovery [52.498055901649025]
Concept Bottleneck Models (CBMs) have been proposed to address the 'black-box' problem of deep neural networks.
We propose a novel CBM approach -- called Discover-then-Name-CBM (DN-CBM) -- that inverts the typical paradigm.
Our concept extraction strategy is efficient, since it is agnostic to the downstream task, and uses concepts already known to the model.
arXiv Detail & Related papers (2024-07-19T17:50:11Z) - Semi-supervised Concept Bottleneck Models [9.875244481114489]
We propose a new framework called SSCBM (Semi-supervised Concept Bottleneck Model)
Our SSCBM is suitable for practical situations where annotated data is scarce.
With only 20% labeled data, we achieved 93.19% concept accuracy and 75.51% (79.82% in a fully supervised setting) prediction accuracy.
arXiv Detail & Related papers (2024-06-27T08:33:35Z) - Sparse Concept Bottleneck Models: Gumbel Tricks in Contrastive Learning [86.15009879251386]
We propose a novel architecture and method of explainable classification with Concept Bottleneck Models (CBM)
CBMs require an additional set of concepts to leverage.
We show a significant increase in accuracy using sparse hidden layers in CLIP-based bottleneck models.
arXiv Detail & Related papers (2024-04-04T09:43:43Z) - Can we Constrain Concept Bottleneck Models to Learn Semantically Meaningful Input Features? [0.6401548653313325]
Concept Bottleneck Models (CBMs) are regarded as inherently interpretable because they first predict a set of human-defined concepts.
Current literature suggests that concept predictions often rely on irrelevant input features.
In this paper, we demonstrate that CBMs can learn to map concepts to semantically meaningful input features.
arXiv Detail & Related papers (2024-02-01T10:18:43Z) - Statistically Significant Concept-based Explanation of Image Classifiers
via Model Knockoffs [22.576922942465142]
Concept-based explanations may cause false positives, which misregards unrelated concepts as important for the prediction task.
We propose a method using a deep learning model to learn the image concept and then using the Knockoff samples to select the important concepts for prediction.
arXiv Detail & Related papers (2023-05-27T05:40:05Z) - Learn to explain yourself, when you can: Equipping Concept Bottleneck
Models with the ability to abstain on their concept predictions [21.94901195358998]
We show how to equip a neural network based classifier with the ability to abstain from predicting concepts when the concept labeling component is uncertain.
Our model learns to provide rationales for its predictions, but only whenever it is sure the rationale is correct.
arXiv Detail & Related papers (2022-11-21T18:07:14Z) - Concept Gradient: Concept-based Interpretation Without Linear Assumption [77.96338722483226]
Concept Activation Vector (CAV) relies on learning a linear relation between some latent representation of a given model and concepts.
We proposed Concept Gradient (CG), extending concept-based interpretation beyond linear concept functions.
We demonstrated CG outperforms CAV in both toy examples and real world datasets.
arXiv Detail & Related papers (2022-08-31T17:06:46Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Concept Bottleneck Models [79.91795150047804]
State-of-the-art models today do not typically support the manipulation of concepts like "the existence of bone spurs"
We revisit the classic idea of first predicting concepts that are provided at training time, and then using these concepts to predict the label.
On x-ray grading and bird identification, concept bottleneck models achieve competitive accuracy with standard end-to-end models.
arXiv Detail & Related papers (2020-07-09T07:47:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.