Multi-Representation Adaptation Network for Cross-domain Image
Classification
- URL: http://arxiv.org/abs/2201.01002v1
- Date: Tue, 4 Jan 2022 06:34:48 GMT
- Title: Multi-Representation Adaptation Network for Cross-domain Image
Classification
- Authors: Yongchun Zhu, Fuzhen Zhuang, Jindong Wang, Jingwu Chen, Zhiping Shi,
Wenjuan Wu, Qing He
- Abstract summary: In image classification, it is often expensive and time-consuming to acquire sufficient labels.
Existing approaches mainly align the distributions of representations extracted by a single structure.
We propose Multi-Representation Adaptation which can dramatically improve the classification accuracy for cross-domain image classification.
- Score: 20.615155915233693
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In image classification, it is often expensive and time-consuming to acquire
sufficient labels. To solve this problem, domain adaptation often provides an
attractive option given a large amount of labeled data from a similar nature
but different domain. Existing approaches mainly align the distributions of
representations extracted by a single structure and the representations may
only contain partial information, e.g., only contain part of the saturation,
brightness, and hue information. Along this line, we propose
Multi-Representation Adaptation which can dramatically improve the
classification accuracy for cross-domain image classification and specially
aims to align the distributions of multiple representations extracted by a
hybrid structure named Inception Adaptation Module (IAM). Based on this, we
present Multi-Representation Adaptation Network (MRAN) to accomplish the
cross-domain image classification task via multi-representation alignment which
can capture the information from different aspects. In addition, we extend
Maximum Mean Discrepancy (MMD) to compute the adaptation loss. Our approach can
be easily implemented by extending most feed-forward models with IAM, and the
network can be trained efficiently via back-propagation. Experiments conducted
on three benchmark image datasets demonstrate the effectiveness of MRAN. The
code has been available at https://github.com/easezyc/deep-transfer-learning.
Related papers
- Fine-grained Recognition with Learnable Semantic Data Augmentation [68.48892326854494]
Fine-grained image recognition is a longstanding computer vision challenge.
We propose diversifying the training data at the feature-level to alleviate the discriminative region loss problem.
Our method significantly improves the generalization performance on several popular classification networks.
arXiv Detail & Related papers (2023-09-01T11:15:50Z) - Interpretable Small Training Set Image Segmentation Network Originated
from Multi-Grid Variational Model [5.283735137946097]
Deep learning (DL) methods have been proposed and widely used for image segmentation.
DL methods usually require a large amount of manually segmented data as training data and suffer from poor interpretability.
In this paper, we replace the hand-crafted regularity term in the MS model with a data adaptive generalized learnable regularity term.
arXiv Detail & Related papers (2023-06-25T02:34:34Z) - UniDAformer: Unified Domain Adaptive Panoptic Segmentation Transformer
via Hierarchical Mask Calibration [49.16591283724376]
We design UniDAformer, a unified domain adaptive panoptic segmentation transformer that is simple but can achieve domain adaptive instance segmentation and semantic segmentation simultaneously within a single network.
UniDAformer introduces Hierarchical Mask (HMC) that rectifies inaccurate predictions at the level of regions, superpixels and annotated pixels via online self-training on the fly.
It has three unique features: 1) it enables unified domain adaptive panoptic adaptation; 2) it mitigates false predictions and improves domain adaptive panoptic segmentation effectively; 3) it is end-to-end trainable with a much simpler training and inference pipeline.
arXiv Detail & Related papers (2022-06-30T07:32:23Z) - AF$_2$: Adaptive Focus Framework for Aerial Imagery Segmentation [86.44683367028914]
Aerial imagery segmentation has some unique challenges, the most critical one among which lies in foreground-background imbalance.
We propose Adaptive Focus Framework (AF$), which adopts a hierarchical segmentation procedure and focuses on adaptively utilizing multi-scale representations.
AF$ has significantly improved the accuracy on three widely used aerial benchmarks, as fast as the mainstream method.
arXiv Detail & Related papers (2022-02-18T10:14:45Z) - Domain Adaptor Networks for Hyperspectral Image Recognition [35.95313368586933]
We consider the problem of adapting a network trained on three-channel color images to a hyperspectral domain with a large number of channels.
We propose domain adaptor networks that map the input to be compatible with a network trained on large-scale color image datasets such as ImageNet.
arXiv Detail & Related papers (2021-08-03T15:06:39Z) - Multi-dataset Pretraining: A Unified Model for Semantic Segmentation [97.61605021985062]
We propose a unified framework, termed as Multi-Dataset Pretraining, to take full advantage of the fragmented annotations of different datasets.
This is achieved by first pretraining the network via the proposed pixel-to-prototype contrastive loss over multiple datasets.
In order to better model the relationship among images and classes from different datasets, we extend the pixel level embeddings via cross dataset mixing.
arXiv Detail & Related papers (2021-06-08T06:13:11Z) - Semantic Distribution-aware Contrastive Adaptation for Semantic
Segmentation [50.621269117524925]
Domain adaptive semantic segmentation refers to making predictions on a certain target domain with only annotations of a specific source domain.
We present a semantic distribution-aware contrastive adaptation algorithm that enables pixel-wise representation alignment.
We evaluate SDCA on multiple benchmarks, achieving considerable improvements over existing algorithms.
arXiv Detail & Related papers (2021-05-11T13:21:25Z) - A Universal Representation Transformer Layer for Few-Shot Image
Classification [43.31379752656756]
Few-shot classification aims to recognize unseen classes when presented with only a small number of samples.
We consider the problem of multi-domain few-shot image classification, where unseen classes and examples come from diverse data sources.
Here, we propose a Universal Representation Transformer layer, that meta-learns to leverage universal features for few-shot classification.
arXiv Detail & Related papers (2020-06-21T03:08:00Z) - Split-Merge Pooling [36.2980225204665]
Split-Merge pooling is introduced to preserve spatial information without subsampling.
We evaluate our approach for dense semantic segmentation of large image sizes taken from the Cityscapes and GTA-5 datasets.
arXiv Detail & Related papers (2020-06-13T23:20:30Z) - Selecting Relevant Features from a Multi-domain Representation for
Few-shot Classification [91.67977602992657]
We propose a new strategy based on feature selection, which is both simpler and more effective than previous feature adaptation approaches.
We show that a simple non-parametric classifier built on top of such features produces high accuracy and generalizes to domains never seen during training.
arXiv Detail & Related papers (2020-03-20T15:44:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.