Slimmable Domain Adaptation
- URL: http://arxiv.org/abs/2206.06620v1
- Date: Tue, 14 Jun 2022 06:28:04 GMT
- Title: Slimmable Domain Adaptation
- Authors: Rang Meng, Weijie Chen, Shicai Yang, Jie Song, Luojun Lin, Di Xie,
Shiliang Pu, Xinchao Wang, Mingli Song, Yueting Zhuang
- Abstract summary: We introduce a simple framework, Slimmable Domain Adaptation, to improve cross-domain generalization with a weight-sharing model bank.
Our framework surpasses other competing approaches by a very large margin on multiple benchmarks.
- Score: 112.19652651687402
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vanilla unsupervised domain adaptation methods tend to optimize the model
with fixed neural architecture, which is not very practical in real-world
scenarios since the target data is usually processed by different
resource-limited devices. It is therefore of great necessity to facilitate
architecture adaptation across various devices. In this paper, we introduce a
simple framework, Slimmable Domain Adaptation, to improve cross-domain
generalization with a weight-sharing model bank, from which models of different
capacities can be sampled to accommodate different accuracy-efficiency
trade-offs. The main challenge in this framework lies in simultaneously
boosting the adaptation performance of numerous models in the model bank. To
tackle this problem, we develop a Stochastic EnsEmble Distillation method to
fully exploit the complementary knowledge in the model bank for inter-model
interaction. Nevertheless, considering the optimization conflict between
inter-model interaction and intra-model adaptation, we augment the existing
bi-classifier domain confusion architecture into an Optimization-Separated
Tri-Classifier counterpart. After optimizing the model bank, architecture
adaptation is leveraged via our proposed Unsupervised Performance Evaluation
Metric. Under various resource constraints, our framework surpasses other
competing approaches by a very large margin on multiple benchmarks. It is also
worth emphasizing that our framework can preserve the performance improvement
against the source-only model even when the computing complexity is reduced to
$1/64$. Code will be available at https://github.com/hikvision-research/SlimDA.
Related papers
- HM3: Hierarchical Multi-Objective Model Merging for Pretrained Models [28.993221775758702]
Model merging is a technique that combines multiple large pretrained models into a single model with enhanced performance and broader task adaptability.
This paper marks a significant advance toward more flexible and comprehensive model merging techniques.
We train policy and value networks using offline sampling of weight vectors, which are then employed for the online optimization of merging strategies.
arXiv Detail & Related papers (2024-09-27T16:31:31Z) - Solving Inverse Problems with Model Mismatch using Untrained Neural Networks within Model-based Architectures [14.551812310439004]
We introduce an untrained forward model residual block within the model-based architecture to match the data consistency in the measurement domain for each instance.
Our approach offers a unified solution that is less parameter-sensitive, requires no additional data, and enables simultaneous fitting of the forward model and reconstruction in a single pass.
arXiv Detail & Related papers (2024-03-07T19:02:13Z) - Variational Model Perturbation for Source-Free Domain Adaptation [64.98560348412518]
We introduce perturbations into the model parameters by variational Bayesian inference in a probabilistic framework.
We demonstrate the theoretical connection to learning Bayesian neural networks, which proves the generalizability of the perturbed model to target domains.
arXiv Detail & Related papers (2022-10-19T08:41:19Z) - Semi-Supervised Domain Adaptation with Auto-Encoder via Simultaneous
Learning [18.601226898819476]
We present a new semi-supervised domain adaptation framework that combines a novel auto-encoder-based domain adaptation model with a simultaneous learning scheme.
Our framework holds strong distribution matching property by training both source and target auto-encoders.
arXiv Detail & Related papers (2022-10-18T00:10:11Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - SlimSeg: Slimmable Semantic Segmentation with Boundary Supervision [54.16430358203348]
We propose a simple but effective slimmable semantic segmentation (SlimSeg) method, which can be executed at different capacities during inference.
We show that our proposed SlimSeg with various mainstream networks can produce flexible models that provide dynamic adjustment of computational cost and better performance.
arXiv Detail & Related papers (2022-07-13T14:41:05Z) - Data Summarization via Bilevel Optimization [48.89977988203108]
A simple yet powerful approach is to operate on small subsets of data.
In this work, we propose a generic coreset framework that formulates the coreset selection as a cardinality-constrained bilevel optimization problem.
arXiv Detail & Related papers (2021-09-26T09:08:38Z) - AutoAdapt: Automated Segmentation Network Search for Unsupervised Domain
Adaptation [4.793219747021116]
We perform neural architecture search (NAS) to provide architecture-level perspective and analysis for domain adaptation.
We propose bridging this gap by using maximum mean discrepancy and regional weighted entropy to estimate the accuracy metric.
arXiv Detail & Related papers (2021-06-24T17:59:02Z) - Semantic Correspondence with Transformers [68.37049687360705]
We propose Cost Aggregation with Transformers (CATs) to find dense correspondences between semantically similar images.
We include appearance affinity modelling to disambiguate the initial correlation maps and multi-level aggregation.
We conduct experiments to demonstrate the effectiveness of the proposed model over the latest methods and provide extensive ablation studies.
arXiv Detail & Related papers (2021-06-04T14:39:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.