Improving Predictive Performance and Calibration by Weight Fusion in
Semantic Segmentation
- URL: http://arxiv.org/abs/2207.11211v1
- Date: Fri, 22 Jul 2022 17:24:13 GMT
- Title: Improving Predictive Performance and Calibration by Weight Fusion in
Semantic Segmentation
- Authors: Timo S\"amann, Ahmed Mostafa Hammam, Andrei Bursuc, Christoph Stiller,
Horst-Michael Gro{\ss}
- Abstract summary: Averaging predictions of a deep ensemble of networks is a popular and effective method to improve predictive performance andcalibration.
We show that a simple weight fusion (WF)strategy can lead to a significantly improved predictive performance andcalibration.
- Score: 18.47581580698701
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Averaging predictions of a deep ensemble of networks is apopular and
effective method to improve predictive performance andcalibration in various
benchmarks and Kaggle competitions. However, theruntime and training cost of
deep ensembles grow linearly with the size ofthe ensemble, making them
unsuitable for many applications. Averagingensemble weights instead of
predictions circumvents this disadvantageduring inference and is typically
applied to intermediate checkpoints ofa model to reduce training cost. Albeit
effective, only few works haveimproved the understanding and the performance of
weight averaging.Here, we revisit this approach and show that a simple weight
fusion (WF)strategy can lead to a significantly improved predictive performance
andcalibration. We describe what prerequisites the weights must meet interms of
weight space, functional space and loss. Furthermore, we presenta new test
method (called oracle test) to measure the functional spacebetween weights. We
demonstrate the versatility of our WF strategy acrossstate of the art
segmentation CNNs and Transformers as well as real worlddatasets such as
BDD100K and Cityscapes. We compare WF with similarapproaches and show our
superiority for in- and out-of-distribution datain terms of predictive
performance and calibration.
Related papers
- Optimizing importance weighting in the presence of sub-population shifts [0.0]
A distribution shift between the training and test data can severely harm performance of machine learning models.
We argue that existing weightings for determining the weights are suboptimal, as they neglect the increase of the variance of the estimated model due to the finite sample size of the training data.
We propose a bi-level optimization procedure in which the weights and model parameters are optimized simultaneously.
arXiv Detail & Related papers (2024-10-18T09:21:10Z) - Forecast-PEFT: Parameter-Efficient Fine-Tuning for Pre-trained Motion Forecasting Models [68.23649978697027]
Forecast-PEFT is a fine-tuning strategy that freezes the majority of the model's parameters, focusing adjustments on newly introduced prompts and adapters.
Our experiments show that Forecast-PEFT outperforms traditional full fine-tuning methods in motion prediction tasks.
Forecast-FT further improves prediction performance, evidencing up to a 9.6% enhancement over conventional baseline methods.
arXiv Detail & Related papers (2024-07-28T19:18:59Z) - Rethinking the Starting Point: Collaborative Pre-Training for Federated Downstream Tasks [21.842345900168525]
CoPreFL is a model-agnostic meta-learning (MAML) procedure that tailors the global model to closely mimic heterogeneous and unseen FL scenarios.
Our MAML procedure incorporates performance variance into the meta-objective function, balancing performance across clients.
We demonstrate that CoPreFL obtains significant improvements in both average accuracy and variance across arbitrary downstream FL tasks.
arXiv Detail & Related papers (2024-02-03T17:58:43Z) - Boosting Fair Classifier Generalization through Adaptive Priority Reweighing [59.801444556074394]
A performance-promising fair algorithm with better generalizability is needed.
This paper proposes a novel adaptive reweighing method to eliminate the impact of the distribution shifts between training and test data on model generalizability.
arXiv Detail & Related papers (2023-09-15T13:04:55Z) - Fast as CHITA: Neural Network Pruning with Combinatorial Optimization [9.440450886684603]
We propose a novel optimization-based pruning framework that considers the combined effect of pruning (and updating) multiple weights subject to a sparsity constraint.
Our approach, CHITA, extends the classical Brain Surgeon framework and results in significant improvements in speed, memory, and performance.
arXiv Detail & Related papers (2023-02-28T15:03:18Z) - Leveraging Angular Information Between Feature and Classifier for
Long-tailed Learning: A Prediction Reformulation Approach [90.77858044524544]
We reformulate the recognition probabilities through included angles without re-balancing the classifier weights.
Inspired by the performance improvement of the predictive form reformulation, we explore the different properties of this angular prediction.
Our method is able to obtain the best performance among peer methods without pretraining on CIFAR10/100-LT and ImageNet-LT.
arXiv Detail & Related papers (2022-12-03T07:52:48Z) - Generalized Parametric Contrastive Learning [60.62901294843829]
Generalized Parametric Contrastive Learning (GPaCo/PaCo) works well on both imbalanced and balanced data.
Experiments on long-tailed benchmarks manifest the new state-of-the-art for long-tailed recognition.
arXiv Detail & Related papers (2022-09-26T03:49:28Z) - Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification [74.62203971625173]
Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
arXiv Detail & Related papers (2022-08-05T01:23:54Z) - FORML: Learning to Reweight Data for Fairness [2.105564340986074]
We introduce Fairness Optimized Reweighting via Meta-Learning (FORML)
FORML balances fairness constraints and accuracy by jointly optimizing training sample weights and a neural network's parameters.
We show that FORML improves equality of opportunity fairness criteria over existing state-of-the-art reweighting methods by approximately 1% on image classification tasks and by approximately 5% on a face prediction task.
arXiv Detail & Related papers (2022-02-03T17:36:07Z) - Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.