Joint Progressive Knowledge Distillation and Unsupervised Domain
Adaptation
- URL: http://arxiv.org/abs/2005.07839v1
- Date: Sat, 16 May 2020 01:07:03 GMT
- Title: Joint Progressive Knowledge Distillation and Unsupervised Domain
Adaptation
- Authors: Le Thanh Nguyen-Meidine, Eric Granger, Madhu Kiran, Jose Dolz,
Louis-Antoine Blais-Morin
- Abstract summary: We propose an unexplored direction -- the joint optimization of CNNs to provide a compressed model that is adapted to perform well for a given target domain.
Our method is compared against state-of-the-art compression and UDA techniques, using two popular classification datasets for UDA -- Office31 and ImageClef-DA.
- Score: 15.115086812609182
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Currently, the divergence in distributions of design and operational data,
and large computational complexity are limiting factors in the adoption of CNNs
in real-world applications. For instance, person re-identification systems
typically rely on a distributed set of cameras, where each camera has different
capture conditions. This can translate to a considerable shift between source
(e.g. lab setting) and target (e.g. operational camera) domains. Given the cost
of annotating image data captured for fine-tuning in each target domain,
unsupervised domain adaptation (UDA) has become a popular approach to adapt
CNNs. Moreover, state-of-the-art deep learning models that provide a high level
of accuracy often rely on architectures that are too complex for real-time
applications. Although several compression and UDA approaches have recently
been proposed to overcome these limitations, they do not allow optimizing a CNN
to simultaneously address both. In this paper, we propose an unexplored
direction -- the joint optimization of CNNs to provide a compressed model that
is adapted to perform well for a given target domain. In particular, the
proposed approach performs unsupervised knowledge distillation (KD) from a
complex teacher model to a compact student model, by leveraging both source and
target data. It also improves upon existing UDA techniques by progressively
teaching the student about domain-invariant features, instead of directly
adapting a compact model on target domain data. Our method is compared against
state-of-the-art compression and UDA techniques, using two popular
classification datasets for UDA -- Office31 and ImageClef-DA. In both datasets,
results indicate that our method can achieve the highest level of accuracy
while requiring a comparable or lower time complexity.
Related papers
- DG-TTA: Out-of-domain medical image segmentation through Domain Generalization and Test-Time Adaptation [43.842694540544194]
We propose to combine domain generalization and test-time adaptation to create a highly effective approach for reusing pre-trained models in unseen target domains.
We demonstrate that our method, combined with pre-trained whole-body CT models, can effectively segment MR images with high accuracy.
arXiv Detail & Related papers (2023-12-11T10:26:21Z) - Informative Data Mining for One-Shot Cross-Domain Semantic Segmentation [84.82153655786183]
We propose a novel framework called Informative Data Mining (IDM) to enable efficient one-shot domain adaptation for semantic segmentation.
IDM provides an uncertainty-based selection criterion to identify the most informative samples, which facilitates quick adaptation and reduces redundant training.
Our approach outperforms existing methods and achieves a new state-of-the-art one-shot performance of 56.7%/55.4% on the GTA5/SYNTHIA to Cityscapes adaptation tasks.
arXiv Detail & Related papers (2023-09-25T15:56:01Z) - CNN Feature Map Augmentation for Single-Source Domain Generalization [6.053629733936548]
Domain Generalization (DG) has gained significant traction during the past few years.
The goal in DG is to produce models which continue to perform well when presented with data distributions different from the ones available during training.
We propose an alternative regularization technique for convolutional neural network architectures in the single-source DG image classification setting.
arXiv Detail & Related papers (2023-05-26T08:48:17Z) - Deep Unsupervised Domain Adaptation: A Review of Recent Advances and
Perspectives [16.68091981866261]
Unsupervised domain adaptation (UDA) is proposed to counter the performance drop on data in a target domain.
UDA has yielded promising results on natural image processing, video analysis, natural language processing, time-series data analysis, medical image analysis, etc.
arXiv Detail & Related papers (2022-08-15T20:05:07Z) - Back to the Source: Diffusion-Driven Test-Time Adaptation [77.4229736436935]
Test-time adaptation harnesses test inputs to improve accuracy of a model trained on source data when tested on shifted target data.
We instead update the target data, by projecting all test inputs toward the source domain with a generative diffusion model.
arXiv Detail & Related papers (2022-07-07T17:14:10Z) - Adapting the Mean Teacher for keypoint-based lung registration under
geometric domain shifts [75.51482952586773]
deep neural networks generally require plenty of labeled training data and are vulnerable to domain shifts between training and test data.
We present a novel approach to geometric domain adaptation for image registration, adapting a model from a labeled source to an unlabeled target domain.
Our method consistently improves on the baseline model by 50%/47% while even matching the accuracy of models trained on target data.
arXiv Detail & Related papers (2022-07-01T12:16:42Z) - Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training
for Road Segmentation of Remote Sensing Images [93.50240389540252]
Road segmentation from remote sensing images is a challenging task with wide ranges of application potentials.
We propose a novel stagewise domain adaptation model called RoadDA to address the domain shift (DS) issue in this field.
Experiment results on two benchmarks demonstrate that RoadDA can efficiently reduce the domain gap and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2021-08-28T09:29:14Z) - Unsupervised and self-adaptative techniques for cross-domain person
re-identification [82.54691433502335]
Person Re-Identification (ReID) across non-overlapping cameras is a challenging task.
Unsupervised Domain Adaptation (UDA) is a promising alternative, as it performs feature-learning adaptation from a model trained on a source to a target domain without identity-label annotation.
In this paper, we propose a novel UDA-based ReID method that takes advantage of triplets of samples created by a new offline strategy.
arXiv Detail & Related papers (2021-03-21T23:58:39Z) - Knowledge Distillation Methods for Efficient Unsupervised Adaptation
Across Multiple Domains [13.464493273131591]
We propose a progressive KD approach for unsupervised single-target DA (STDA) and multi-target DA (MTDA) of CNNs.
Our proposed approach is compared against state-of-the-art methods for compression and STDA of CNNs on the Office31 and ImageClef-DA image classification datasets.
arXiv Detail & Related papers (2021-01-18T19:53:16Z) - Keep it Simple: Image Statistics Matching for Domain Adaptation [0.0]
Domain Adaptation (DA) is a technique to maintain detection accuracy when only unlabeled images are available of the target domain.
Recent state-of-the-art methods try to reduce the domain gap using an adversarial training strategy.
We propose to align either color histograms or mean and covariance of the source images towards the target domain.
In comparison to recent methods, we achieve state-of-the-art performance using a much simpler procedure for the training.
arXiv Detail & Related papers (2020-05-26T07:32:09Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.