Synchronizing Task Behavior: Aligning Multiple Tasks during Test-Time Training
- URL: http://arxiv.org/abs/2507.07778v2
- Date: Mon, 21 Jul 2025 05:48:40 GMT
- Title: Synchronizing Task Behavior: Aligning Multiple Tasks during Test-Time Training
- Authors: Wooseong Jeong, Jegyeong Cho, Youngho Yoon, Kuk-Jin Yoon,
- Abstract summary: Generalizing neural networks to unseen target domains is a significant challenge in real-world deployments.<n>Test-time training (TTT) addresses this by using an auxiliary self-supervised task to reduce the domain gap caused by distribution shifts between the source and target.<n>We propose a novel TTT approach called Synchronizing Tasks for Test-time Training (S4T), which enables the concurrent handling of multiple tasks.
- Score: 39.2014649652258
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generalizing neural networks to unseen target domains is a significant challenge in real-world deployments. Test-time training (TTT) addresses this by using an auxiliary self-supervised task to reduce the domain gap caused by distribution shifts between the source and target. However, we find that when models are required to perform multiple tasks under domain shifts, conventional TTT methods suffer from unsynchronized task behavior, where the adaptation steps needed for optimal performance in one task may not align with the requirements of other tasks. To address this, we propose a novel TTT approach called Synchronizing Tasks for Test-time Training (S4T), which enables the concurrent handling of multiple tasks. The core idea behind S4T is that predicting task relations across domain shifts is key to synchronizing tasks during test time. To validate our approach, we apply S4T to conventional multi-task benchmarks, integrating it with traditional TTT protocols. Our empirical results show that S4T outperforms state-of-the-art TTT methods across various benchmarks.
Related papers
- Dynamic Task Vector Grouping for Efficient Multi-Task Prompt Tuning [20.37803751979975]
Multi-task prompt tuning utilizes multiple high-resource source tasks to improve performance on low-source target tasks.<n>Existing approaches transfer the soft prompt trained by combining all source tasks or a single high-similar'' source task one-time-only iteration.<n>We find that the optimal transfer performance often comes from a combination of source tasks, which is neither one nor all.
arXiv Detail & Related papers (2025-03-23T13:09:04Z) - MT2ST: Adaptive Multi-Task to Single-Task Learning [6.185573921868495]
We address the trade-off between generalization in multi-task learning (MTL) and precision in single-task learning (STL) by introducing the Multi-Task to Single-Task (MT2ST) framework.<n>MT2ST is designed to enhance training efficiency and accuracy in multi-modal tasks, showcasing its value as a practical application of efficient machine learning.
arXiv Detail & Related papers (2024-06-26T03:12:07Z) - ClusT3: Information Invariant Test-Time Training [19.461441044484427]
Test-time training (TTT) methods have been developed in an attempt to mitigate these vulnerabilities.
We propose a novel unsupervised TTT technique based on the Mutual of Mutual Information between multi-scale feature maps and a discrete latent representation.
Experimental results demonstrate competitive classification performance on different popular test-time adaptation benchmarks.
arXiv Detail & Related papers (2023-10-18T21:43:37Z) - Improved Test-Time Adaptation for Domain Generalization [48.239665441875374]
Test-time training (TTT) adapts the learned model with test data.
This work addresses two main factors: selecting an appropriate auxiliary TTT task for updating and identifying reliable parameters to update during the test phase.
We introduce additional adaptive parameters for the trained model, and we suggest only updating the adaptive parameters during the test phase.
arXiv Detail & Related papers (2023-04-10T10:12:38Z) - Revisiting Realistic Test-Time Training: Sequential Inference and
Adaptation by Anchored Clustering Regularized Self-Training [37.75537703971045]
We develop a test-time anchored clustering (TTAC) approach to enable stronger test-time feature learning.
Self-training(ST) has demonstrated great success in learning from unlabeled data.
TTAC++ consistently outperforms the state-of-the-art methods on five TTT datasets.
arXiv Detail & Related papers (2023-03-20T04:30:18Z) - Frustratingly Simple Entity Tracking with Effective Use of Multi-Task
Learning Models [5.9585526937249]
We present SET, a frustratingly Simple-yet-effective approach for entity tracking in procedural text.
Compared with state-of-the-art entity tracking models that require domain-specific pre-training, SET simply fine-tunes off-the-shelf T5 with customized formats.
We show that T5's supervised multi-task learning plays an important role in the success of SET.
arXiv Detail & Related papers (2022-10-12T17:46:16Z) - Revisiting Realistic Test-Time Training: Sequential Inference and
Adaptation by Anchored Clustering [37.76664203157892]
We develop a test-time anchored clustering (TTAC) approach to enable stronger test-time feature learning.
TTAC discovers clusters in both source and target domain and match the target clusters to the source ones to improve generalization.
We demonstrate that under all TTT protocols TTAC consistently outperforms the state-of-the-art methods on five TTT datasets.
arXiv Detail & Related papers (2022-06-06T16:23:05Z) - FETA: A Benchmark for Few-Sample Task Transfer in Open-Domain Dialogue [70.65782786401257]
This work explores conversational task transfer by introducing FETA: a benchmark for few-sample task transfer in open-domain dialogue.
FETA contains two underlying sets of conversations upon which there are 10 and 7 tasks annotated, enabling the study of intra-dataset task transfer.
We utilize three popular language models and three learning algorithms to analyze the transferability between 132 source-target task pairs.
arXiv Detail & Related papers (2022-05-12T17:59:00Z) - Task Adaptive Parameter Sharing for Multi-Task Learning [114.80350786535952]
Adaptive Task Adapting Sharing (TAPS) is a method for tuning a base model to a new task by adaptively modifying a small, task-specific subset of layers.
Compared to other methods, TAPS retains high accuracy on downstream tasks while introducing few task-specific parameters.
We evaluate our method on a suite of fine-tuning tasks and architectures (ResNet, DenseNet, ViT) and show that it achieves state-of-the-art performance while being simple to implement.
arXiv Detail & Related papers (2022-03-30T23:16:07Z) - Weighted Training for Cross-Task Learning [71.94908559469475]
We introduce Target-Aware Weighted Training (TAWT), a weighted training algorithm for cross-task learning.
We show that TAWT is easy to implement, is computationally efficient, requires little hyper parameter tuning, and enjoys non-asymptotic learning-theoretic guarantees.
As a byproduct, the proposed representation-based task distance allows one to reason in a theoretically principled way about several critical aspects of cross-task learning.
arXiv Detail & Related papers (2021-05-28T20:27:02Z) - Exploring Relational Context for Multi-Task Dense Prediction [76.86090370115]
We consider a multi-task environment for dense prediction tasks, represented by a common backbone and independent task-specific heads.
We explore various attention-based contexts, such as global and local, in the multi-task setting.
We propose an Adaptive Task-Relational Context module, which samples the pool of all available contexts for each task pair.
arXiv Detail & Related papers (2021-04-28T16:45:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.