Learning from Extrinsic and Intrinsic Supervisions for Domain
Generalization
- URL: http://arxiv.org/abs/2007.09316v1
- Date: Sat, 18 Jul 2020 03:12:24 GMT
- Title: Learning from Extrinsic and Intrinsic Supervisions for Domain
Generalization
- Authors: Shujun Wang, Lequan Yu, Caizi Li, Chi-Wing Fu, and Pheng-Ann Heng
- Abstract summary: We present a new domain generalization framework that learns how to generalize across domains simultaneously.
We demonstrate the effectiveness of our approach on two standard object recognition benchmarks.
- Score: 95.73898853032865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The generalization capability of neural networks across domains is crucial
for real-world applications. We argue that a generalized object recognition
system should well understand the relationships among different images and also
the images themselves at the same time. To this end, we present a new domain
generalization framework that learns how to generalize across domains
simultaneously from extrinsic relationship supervision and intrinsic
self-supervision for images from multi-source domains. To be specific, we
formulate our framework with feature embedding using a multi-task learning
paradigm. Besides conducting the common supervised recognition task, we
seamlessly integrate a momentum metric learning task and a self-supervised
auxiliary task to collectively utilize the extrinsic supervision and intrinsic
supervision. Also, we develop an effective momentum metric learning scheme with
K-hard negative mining to boost the network to capture image relationship for
domain generalization. We demonstrate the effectiveness of our approach on two
standard object recognition benchmarks VLCS and PACS, and show that our methods
achieve state-of-the-art performance.
Related papers
- Multi-Scale and Multi-Layer Contrastive Learning for Domain Generalization [5.124256074746721]
We argue that the generalization ability of deep convolutional neural networks can be improved by taking advantage of multi-layer and multi-scaled representations of the network.
We introduce a framework that aims at improving domain generalization of image classifiers by combining both low-level and high-level features at multiple scales.
We show that our model is able to surpass the performance of previous DG methods and consistently produce competitive and state-of-the-art results in all datasets.
arXiv Detail & Related papers (2023-08-28T08:54:27Z) - Modeling Multiple Views via Implicitly Preserving Global Consistency and
Local Complementarity [61.05259660910437]
We propose a global consistency and complementarity network (CoCoNet) to learn representations from multiple views.
On the global stage, we reckon that the crucial knowledge is implicitly shared among views, and enhancing the encoder to capture such knowledge can improve the discriminability of the learned representations.
Lastly on the local stage, we propose a complementarity-factor, which joints cross-view discriminative knowledge, and it guides the encoders to learn not only view-wise discriminability but also cross-view complementary information.
arXiv Detail & Related papers (2022-09-16T09:24:00Z) - Unsupervised Domain Generalization by Learning a Bridge Across Domains [78.855606355957]
Unsupervised Domain Generalization (UDG) setup has no training supervision in neither source nor target domains.
Our approach is based on self-supervised learning of a Bridge Across Domains (BrAD) - an auxiliary bridge domain accompanied by a set of semantics preserving visual (image-to-image) mappings to BrAD from each of the training domains.
We show how using an edge-regularized BrAD our approach achieves significant gains across multiple benchmarks and a range of tasks, including UDG, Few-shot UDA, and unsupervised generalization across multi-domain datasets.
arXiv Detail & Related papers (2021-12-04T10:25:45Z) - Joint Learning of Neural Transfer and Architecture Adaptation for Image
Recognition [77.95361323613147]
Current state-of-the-art visual recognition systems rely on pretraining a neural network on a large-scale dataset and finetuning the network weights on a smaller dataset.
In this work, we prove that dynamically adapting network architectures tailored for each domain task along with weight finetuning benefits in both efficiency and effectiveness.
Our method can be easily generalized to an unsupervised paradigm by replacing supernet training with self-supervised learning in the source domain tasks and performing linear evaluation in the downstream tasks.
arXiv Detail & Related papers (2021-03-31T08:15:17Z) - Domain-Robust Visual Imitation Learning with Mutual Information
Constraints [0.0]
We introduce a new algorithm called Disentangling Generative Adversarial Imitation Learning (DisentanGAIL)
Our algorithm enables autonomous agents to learn directly from high dimensional observations of an expert performing a task.
arXiv Detail & Related papers (2021-03-08T21:18:58Z) - Self-Supervised Learning Across Domains [33.86614301708017]
We propose to apply a similar approach to the problem of object recognition across domains.
Our model learns the semantic labels in a supervised fashion, and broadens its understanding of the data by learning from self-supervised signals on the same images.
This secondary task helps the network to focus on object shapes, learning concepts like spatial orientation and part correlation, while acting as a regularizer for the classification task.
arXiv Detail & Related papers (2020-07-24T06:19:53Z) - Self-supervised Learning from a Multi-view Perspective [121.63655399591681]
We show that self-supervised representations can extract task-relevant information and discard task-irrelevant information.
Our theoretical framework paves the way to a larger space of self-supervised learning objective design.
arXiv Detail & Related papers (2020-06-10T00:21:35Z) - Improving out-of-distribution generalization via multi-task
self-supervised pretraining [48.29123326140466]
We show that features obtained using self-supervised learning are comparable to, or better than, supervised learning for domain generalization in computer vision.
We introduce a new self-supervised pretext task of predicting responses to Gabor filter banks.
arXiv Detail & Related papers (2020-03-30T14:55:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.