Knowledge Distillation to Ensemble Global and Interpretable
Prototype-Based Mammogram Classification Models
- URL: http://arxiv.org/abs/2209.12420v1
- Date: Mon, 26 Sep 2022 05:04:15 GMT
- Title: Knowledge Distillation to Ensemble Global and Interpretable
Prototype-Based Mammogram Classification Models
- Authors: Chong Wang, Yuanhong Chen, Yuyuan Liu, Yu Tian, Fengbei Liu, Davis J.
McCarthy, Michael Elliott, Helen Frazer, Gustavo Carneiro
- Abstract summary: We propose BRAIxProtoPNet++, which adds interpretability to a global model by ensembling it with a prototype-based model.
We show that BRAIxProtoPNet++ has higher classification accuracy than SOTA global and prototype-based models.
- Score: 20.16068689434846
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: State-of-the-art (SOTA) deep learning mammogram classifiers, trained with
weakly-labelled images, often rely on global models that produce predictions
with limited interpretability, which is a key barrier to their successful
translation into clinical practice. On the other hand, prototype-based models
improve interpretability by associating predictions with training image
prototypes, but they are less accurate than global models and their prototypes
tend to have poor diversity. We address these two issues with the proposal of
BRAIxProtoPNet++, which adds interpretability to a global model by ensembling
it with a prototype-based model. BRAIxProtoPNet++ distills the knowledge of the
global model when training the prototype-based model with the goal of
increasing the classification accuracy of the ensemble. Moreover, we propose an
approach to increase prototype diversity by guaranteeing that all prototypes
are associated with different training images. Experiments on weakly-labelled
private and public datasets show that BRAIxProtoPNet++ has higher
classification accuracy than SOTA global and prototype-based models. Using
lesion localisation to assess model interpretability, we show BRAIxProtoPNet++
is more effective than other prototype-based models and post-hoc explanation of
global models. Finally, we show that the diversity of the prototypes learned by
BRAIxProtoPNet++ is superior to SOTA prototype-based approaches.
Related papers
- This Looks Better than That: Better Interpretable Models with ProtoPNeXt [14.28283868577614]
Prototypical-part models are a popular interpretable alternative to black-box deep learning models for computer vision.
We create a new framework for integrating components of prototypical-part models -- ProtoPNeXt.
arXiv Detail & Related papers (2024-06-20T18:54:27Z) - Prototype-based Interpretable Breast Cancer Prediction Models: Analysis and Challenges [2.2194851850560715]
We show the use of PEF-C in the context of breast cancer prediction using mammography.
Existing works on prototype-based models on breast cancer prediction using mammography have focused on improving the classification performance of prototype-based models.
We are the first to evaluate the quality of the mammography prototypes systematically using our PEF-C.
arXiv Detail & Related papers (2024-03-29T16:08:59Z) - Unsupervised Prototype Adapter for Vision-Language Models [29.516767588241724]
We design an unsupervised fine-tuning approach for vision-language models called Unsupervised Prototype Adapter (UP-Adapter)
Specifically, for the unannotated target datasets, we leverage the text-image aligning capability of CLIP to automatically select the most confident samples for each class.
After fine-tuning, the prototype model prediction is combined with the original CLIP's prediction by a residual connection to perform downstream recognition tasks.
arXiv Detail & Related papers (2023-08-22T15:28:49Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Tackling Data Heterogeneity in Federated Learning with Class Prototypes [44.746340839025194]
We propose FedNH, a novel method that improves the local models' performance for both personalization and generalization.
We show that imposing uniformity helps to combat prototype collapse while infusing class semantics improves local models.
arXiv Detail & Related papers (2022-12-06T05:15:38Z) - Investigating Ensemble Methods for Model Robustness Improvement of Text
Classifiers [66.36045164286854]
We analyze a set of existing bias features and demonstrate there is no single model that works best for all the cases.
By choosing an appropriate bias model, we can obtain a better robustness result than baselines with a more sophisticated model design.
arXiv Detail & Related papers (2022-10-28T17:52:10Z) - Composing Ensembles of Pre-trained Models via Iterative Consensus [95.10641301155232]
We propose a unified framework for composing ensembles of different pre-trained models.
We use pre-trained models as "generators" or "scorers" and compose them via closed-loop iterative consensus optimization.
We demonstrate that consensus achieved by an ensemble of scorers outperforms the feedback of a single scorer.
arXiv Detail & Related papers (2022-10-20T18:46:31Z) - Language Models are General-Purpose Interfaces [109.45478241369655]
We propose to use language models as a general-purpose interface to various foundation models.
A collection of pretrained encoders perceive diverse modalities (such as vision, and language)
We propose a semi-causal language modeling objective to jointly pretrain the interface and the modular encoders.
arXiv Detail & Related papers (2022-06-13T17:34:22Z) - BMD: A General Class-balanced Multicentric Dynamic Prototype Strategy
for Source-free Domain Adaptation [74.93176783541332]
Source-free Domain Adaptation (SFDA) aims to adapt a pre-trained source model to the unlabeled target domain without accessing the well-labeled source data.
To make up for the absence of source data, most existing methods introduced feature prototype based pseudo-labeling strategies.
We propose a general class-Balanced Multicentric Dynamic prototype strategy for the SFDA task.
arXiv Detail & Related papers (2022-04-06T13:23:02Z) - Improving Label Quality by Jointly Modeling Items and Annotators [68.8204255655161]
We propose a fully Bayesian framework for learning ground truth labels from noisy annotators.
Our framework ensures scalability by factoring a generative, Bayesian soft clustering model over label distributions into the classic David and Skene joint annotator-data model.
arXiv Detail & Related papers (2021-06-20T02:15:20Z) - PSD2 Explainable AI Model for Credit Scoring [0.0]
The aim of this project is to develop and test advanced analytical methods to improve the prediction accuracy of Credit Risk Models.
The project focuses on applying an explainable machine learning model to bank-related databases.
arXiv Detail & Related papers (2020-11-20T12:12:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.