Perturbation-Based Two-Stage Multi-Domain Active Learning
- URL: http://arxiv.org/abs/2306.10700v1
- Date: Mon, 19 Jun 2023 04:58:32 GMT
- Title: Perturbation-Based Two-Stage Multi-Domain Active Learning
- Authors: Rui He, Zeyu Dai, Shan He, Ke Tang
- Abstract summary: We propose a perturbation-based two-stage multi-domain active learning (P2S-MDAL) method incorporated into the well-regarded ASP-MTL model.
P2S-MDAL involves allocating budgets for domains and establishing regions for diversity selection.
A perturbation metric has been introduced to evaluate the robustness of the shared feature extractor of the model.
- Score: 31.073745612552926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In multi-domain learning (MDL) scenarios, high labeling effort is required
due to the complexity of collecting data from various domains. Active Learning
(AL) presents an encouraging solution to this issue by annotating a smaller
number of highly informative instances, thereby reducing the labeling effort.
Previous research has relied on conventional AL strategies for MDL scenarios,
which underutilize the domain-shared information of each instance during the
selection procedure. To mitigate this issue, we propose a novel
perturbation-based two-stage multi-domain active learning (P2S-MDAL) method
incorporated into the well-regarded ASP-MTL model. Specifically, P2S-MDAL
involves allocating budgets for domains and establishing regions for diversity
selection, which are further used to select the most cross-domain influential
samples in each region. A perturbation metric has been introduced to evaluate
the robustness of the shared feature extractor of the model, facilitating the
identification of potentially cross-domain influential samples. Experiments are
conducted on three real-world datasets, encompassing both texts and images. The
superior performance over conventional AL strategies shows the effectiveness of
the proposed strategy. Additionally, an ablation study has been carried out to
demonstrate the validity of each component. Finally, we outline several
intriguing potential directions for future MDAL research, thus catalyzing the
field's advancement.
Related papers
- Improving Intrusion Detection with Domain-Invariant Representation Learning in Latent Space [4.871119861180455]
We introduce a two-phase representation learning technique using multi-task learning.
We disentangle the latent space by minimizing the mutual information between the prior and latent space.
We assess the model's efficacy across multiple cybersecurity datasets.
arXiv Detail & Related papers (2023-12-28T17:24:13Z) - Benchmarking Multi-Domain Active Learning on Image Classification [16.690755621494215]
We introduce a multi-domain active learning benchmark to bridge the gap between research on single-source data and real-world data.
Our benchmark demonstrates that traditional single-domain active learning strategies are often less effective than random selection in multi-domain scenarios.
Analysis on our benchmark shows that all multi-domain strategies exhibit significant tradeoffs, with no strategy outperforming across all datasets or all metrics.
arXiv Detail & Related papers (2023-12-01T06:11:14Z) - Revisiting the Domain Shift and Sample Uncertainty in Multi-source
Active Domain Transfer [69.82229895838577]
Active Domain Adaptation (ADA) aims to maximally boost model adaptation in a new target domain by actively selecting a limited number of target data to annotate.
This setting neglects the more practical scenario where training data are collected from multiple sources.
This motivates us to target a new and challenging setting of knowledge transfer that extends ADA from a single source domain to multiple source domains.
arXiv Detail & Related papers (2023-11-21T13:12:21Z) - Improving Anomaly Segmentation with Multi-Granularity Cross-Domain
Alignment [17.086123737443714]
Anomaly segmentation plays a pivotal role in identifying atypical objects in images, crucial for hazard detection in autonomous driving systems.
While existing methods demonstrate noteworthy results on synthetic data, they often fail to consider the disparity between synthetic and real-world data domains.
We introduce the Multi-Granularity Cross-Domain Alignment framework, tailored to harmonize features across domains at both the scene and individual sample levels.
arXiv Detail & Related papers (2023-08-16T22:54:49Z) - Dynamic Domain Discrepancy Adjustment for Active Multi-Domain Adaptation [3.367755441623275]
Multi-source unsupervised domain adaptation (MUDA) aims to transfer knowledge from related source domains to an unlabeled target domain.
We propose a novel approach called Dynamic Domain Discrepancy Adjustment for Active Multi-Domain Adaptation (D3AAMDA)
This mechanism controls the alignment level of features between each source domain and the target domain, effectively leveraging the local advantageous feature information within the source domains.
arXiv Detail & Related papers (2023-07-26T09:40:19Z) - Tackling Long-Tailed Category Distribution Under Domain Shifts [50.21255304847395]
Existing approaches cannot handle the scenario where both issues exist.
We designed three novel core functional blocks including Distribution Calibrated Classification Loss, Visual-Semantic Mapping and Semantic-Similarity Guided Augmentation.
Two new datasets were proposed for this problem, named AWA2-LTS and ImageNet-LTS.
arXiv Detail & Related papers (2022-07-20T19:07:46Z) - Consistency and Diversity induced Human Motion Segmentation [231.36289425663702]
We propose a novel Consistency and Diversity induced human Motion (CDMS) algorithm.
Our model factorizes the source and target data into distinct multi-layer feature spaces.
A multi-mutual learning strategy is carried out to reduce the domain gap between the source and target data.
arXiv Detail & Related papers (2022-02-10T06:23:56Z) - Active Learning Over Multiple Domains in Natural Language Tasks [6.8700723936987975]
We survey a variety of techniques in active learning (AL), domain shift detection (DS), and multi-domain sampling.
Among 18 acquisition functions from 4 families of methods, we find H- Divergence methods, and particularly our proposed variant DAL-E, yield effective results.
Our findings yield the first comprehensive analysis of both existing and novel methods for practitioners faced with multi-domain active learning for natural language tasks.
arXiv Detail & Related papers (2022-02-01T07:27:18Z) - Adaptively-Accumulated Knowledge Transfer for Partial Domain Adaptation [66.74638960925854]
Partial domain adaptation (PDA) deals with a realistic and challenging problem when the source domain label space substitutes the target domain.
We propose an Adaptively-Accumulated Knowledge Transfer framework (A$2$KT) to align the relevant categories across two domains.
arXiv Detail & Related papers (2020-08-27T00:53:43Z) - Text Recognition in Real Scenarios with a Few Labeled Samples [55.07859517380136]
Scene text recognition (STR) is still a hot research topic in computer vision field.
This paper proposes a few-shot adversarial sequence domain adaptation (FASDA) approach to build sequence adaptation.
Our approach can maximize the character-level confusion between the source domain and the target domain.
arXiv Detail & Related papers (2020-06-22T13:03:01Z) - Multi-source Domain Adaptation in the Deep Learning Era: A Systematic
Survey [53.656086832255944]
Multi-source domain adaptation (MDA) is a powerful extension in which the labeled data may be collected from multiple sources.
MDA has attracted increasing attention in both academia and industry.
arXiv Detail & Related papers (2020-02-26T08:07:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.