Role Prompting Guided Domain Adaptation with General Capability Preserve
for Large Language Models
- URL: http://arxiv.org/abs/2403.02756v1
- Date: Tue, 5 Mar 2024 08:22:41 GMT
- Title: Role Prompting Guided Domain Adaptation with General Capability Preserve
for Large Language Models
- Authors: Rui Wang, Fei Mi, Yi Chen, Boyang Xue, Hongru Wang, Qi Zhu, Kam-Fai
Wong, Ruifeng Xu
- Abstract summary: When tailored to specific domains, Large Language Models (LLMs) tend to experience catastrophic forgetting.
crafting a versatile model for multiple domains simultaneously often results in a decline in overall performance.
We present the RolE Prompting Guided Multi-Domain Adaptation (REGA) strategy.
- Score: 55.51408151807268
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The growing interest in Large Language Models (LLMs) for specialized
applications has revealed a significant challenge: when tailored to specific
domains, LLMs tend to experience catastrophic forgetting, compromising their
general capabilities and leading to a suboptimal user experience. Additionally,
crafting a versatile model for multiple domains simultaneously often results in
a decline in overall performance due to confusion between domains. In response
to these issues, we present the RolE Prompting Guided Multi-Domain Adaptation
(REGA) strategy. This novel approach effectively manages multi-domain LLM
adaptation through three key components: 1) Self-Distillation constructs and
replays general-domain exemplars to alleviate catastrophic forgetting. 2) Role
Prompting assigns a central prompt to the general domain and a unique role
prompt to each specific domain to minimize inter-domain confusion during
training. 3) Role Integration reuses and integrates a small portion of
domain-specific data to the general-domain data, which are trained under the
guidance of the central prompt. The central prompt is used for a streamlined
inference process, removing the necessity to switch prompts for different
domains. Empirical results demonstrate that REGA effectively alleviates
catastrophic forgetting and inter-domain confusion. This leads to improved
domain-specific performance compared to standard fine-tuned models, while still
preserving robust general capabilities.
Related papers
- More Than Catastrophic Forgetting: Integrating General Capabilities For Domain-Specific LLMs [40.54076184225558]
The performance on general tasks decreases after Large Language Models (LLMs) are fine-tuned on domain-specific tasks, known as Catastrophic Forgetting (CF)
This paper presents a challenge for real application of domain-specific LLMs beyond CF, called General Capabilities Integration (GCI)
The objective of GCI is not merely to retain previously acquired general capabilities alongside new domain knowledge, but to harmonize and utilize both sets of skills in a cohesive manner to enhance performance on domain-specific tasks.
arXiv Detail & Related papers (2024-05-28T05:00:12Z) - Large-Scale Multi-Domain Recommendation: an Automatic Domain Feature Extraction and Personalized Integration Framework [30.46152832695426]
We propose an Automatic Domain Feature Extraction and Personalized Integration (DFEI) framework for the large-scale multi-domain recommendation.
The framework automatically transforms the behavior of each individual user into an aggregation of all user behaviors within the domain, which serves as the domain features.
Experimental results on both public and industrial datasets, consisting of over 20 domains, clearly demonstrate that the proposed framework achieves significantly better performance compared with SOTA baselines.
arXiv Detail & Related papers (2024-04-12T09:57:17Z) - DIGIC: Domain Generalizable Imitation Learning by Causal Discovery [69.13526582209165]
Causality has been combined with machine learning to produce robust representations for domain generalization.
We make a different attempt by leveraging the demonstration data distribution to discover causal features for a domain generalizable policy.
We design a novel framework, called DIGIC, to identify the causal features by finding the direct cause of the expert action from the demonstration data distribution.
arXiv Detail & Related papers (2024-02-29T07:09:01Z) - Domain Generalization for Domain-Linked Classes [8.738092015092207]
In the real-world, classes may often be domain-linked, i.e. expressed only in a specific domain.
We propose a Fair and cONtrastive feature-space regularization algorithm for Domain-linked DG, FOND.
arXiv Detail & Related papers (2023-06-01T16:39:50Z) - Domain generalization Person Re-identification on Attention-aware
multi-operation strategery [8.90472129039969]
Domain generalization person re-identification (DG Re-ID) aims to directly deploy a model trained on the source domain to the unseen target domain with good generalization.
In the existing DG Re-ID methods, invariant operations are effective in extracting domain generalization features.
An Attention-aware Multi-operation Strategery (AMS) for DG Re-ID is proposed to extract more generalized features.
arXiv Detail & Related papers (2022-10-19T09:18:46Z) - Compound Domain Generalization via Meta-Knowledge Encoding [55.22920476224671]
We introduce Style-induced Domain-specific Normalization (SDNorm) to re-normalize the multi-modal underlying distributions.
We harness the prototype representations, the centroids of classes, to perform relational modeling in the embedding space.
Experiments on four standard Domain Generalization benchmarks reveal that COMEN exceeds the state-of-the-art performance without the need of domain supervision.
arXiv Detail & Related papers (2022-03-24T11:54:59Z) - META: Mimicking Embedding via oThers' Aggregation for Generalizable
Person Re-identification [68.39849081353704]
Domain generalizable (DG) person re-identification (ReID) aims to test across unseen domains without access to the target domain data at training time.
This paper presents a new approach called Mimicking Embedding via oThers' Aggregation (META) for DG ReID.
arXiv Detail & Related papers (2021-12-16T08:06:50Z) - TAL: Two-stream Adaptive Learning for Generalizable Person
Re-identification [115.31432027711202]
We argue that both domain-specific and domain-invariant features are crucial for improving the generalization ability of re-id models.
We name two-stream adaptive learning (TAL) to simultaneously model these two kinds of information.
Our framework can be applied to both single-source and multi-source domain generalization tasks.
arXiv Detail & Related papers (2021-11-29T01:27:42Z) - Structured Latent Embeddings for Recognizing Unseen Classes in Unseen
Domains [108.11746235308046]
We propose a novel approach that learns domain-agnostic structured latent embeddings by projecting images from different domains.
Our experiments on the challenging DomainNet and DomainNet-LS benchmarks show the superiority of our approach over existing methods.
arXiv Detail & Related papers (2021-07-12T17:57:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.