Multi-Order Networks for Action Unit Detection
- URL: http://arxiv.org/abs/2202.00446v1
- Date: Tue, 1 Feb 2022 14:58:21 GMT
- Title: Multi-Order Networks for Action Unit Detection
- Authors: Gauthier Tallec, Arnaud Dapogny and Kevin Bailly
- Abstract summary: Multi-Order Network (MONET) is a multi-task learning method with joint task order optimization.
We show that MONET significantly extends state-of-the-art performance in Facial Action Unit detection.
- Score: 7.971065005161565
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep multi-task methods, where several tasks are learned within a single
network, have recently attracted increasing attention. Burning point of this
attention is their capacity to capture inter-task relationships. Current
approaches either only rely on weight sharing, or add explicit dependency
modelling by decomposing the task joint distribution using Bayes chain rule. If
the latter strategy yields comprehensive inter-task relationships modelling, it
requires imposing an arbitrary order into an unordered task set. Most
importantly, this sequence ordering choice has been identified as a critical
source of performance variations. In this paper, we present Multi-Order Network
(MONET), a multi-task learning method with joint task order optimization. MONET
uses a differentiable order selection based on soft order modelling inside
Birkhoff's polytope to jointly learn task-wise recurrent modules with their
optimal chaining order. Furthermore, we introduce warm up and order dropout to
enhance order selection by encouraging order exploration. Experimentally, we
first validate MONET capacity to retrieve the optimal order in a toy
environment. Second, we use an attribute detection scenario to show that MONET
outperforms existing multi-task baselines on a wide range of dependency
settings. Finally, we demonstrate that MONET significantly extends
state-of-the-art performance in Facial Action Unit detection.
Related papers
- SEMINAR: Search Enhanced Multi-modal Interest Network and Approximate Retrieval for Lifelong Sequential Recommendation [16.370075234443245]
We propose a unified lifelong multi-modal sequence model called SEMINAR-Search Enhanced Multi-Modal Interest Network and Approximate Retrieval.
Specifically, a network called Pretraining Search Unit learns the lifelong sequences of multi-modal query-item pairs in a pretraining-finetuning manner.
To accelerate the online retrieval speed of multi-modal embedding, we propose a multi-modal codebook-based product quantization strategy.
arXiv Detail & Related papers (2024-07-15T13:33:30Z) - DMTG: One-Shot Differentiable Multi-Task Grouping [32.72240053032646]
We aim to address Multi-Task Learning (MTL) with a large number of tasks by Multi-Task Grouping (MTG)
We propose to simultaneously identify the best task groups from 2N candidates and train the model weights simultaneously in one-shot, with the high-order task-affinity fully exploited.
arXiv Detail & Related papers (2024-07-06T13:54:00Z) - Task Indicating Transformer for Task-conditional Dense Predictions [16.92067246179703]
We introduce a novel task-conditional framework called Task Indicating Transformer (TIT) to tackle this challenge.
Our approach designs a Mix Task Adapter module within the transformer block, which incorporates a Task Indicating Matrix through matrix decomposition.
We also propose a Task Gate Decoder module that harnesses a Task Indicating Vector and gating mechanism to facilitate adaptive multi-scale feature refinement.
arXiv Detail & Related papers (2024-03-01T07:06:57Z) - Identification of Negative Transfers in Multitask Learning Using
Surrogate Models [29.882265735630046]
Multitask learning is widely used to train a low-resource target task by augmenting it with multiple related source tasks.
A critical problem in multitask learning is identifying subsets of source tasks that would benefit the target task.
We introduce an efficient procedure to address this problem via surrogate modeling.
arXiv Detail & Related papers (2023-03-25T23:16:11Z) - OFASys: A Multi-Modal Multi-Task Learning System for Building Generalist
Models [72.8156832931841]
Generalist models are capable of performing diverse multi-modal tasks in a task-agnostic way within a single model.
We release a generalist model learning system, OFASys, built on top of a declarative task interface named multi-modal instruction.
arXiv Detail & Related papers (2022-12-08T17:07:09Z) - Task Adaptive Parameter Sharing for Multi-Task Learning [114.80350786535952]
Adaptive Task Adapting Sharing (TAPS) is a method for tuning a base model to a new task by adaptively modifying a small, task-specific subset of layers.
Compared to other methods, TAPS retains high accuracy on downstream tasks while introducing few task-specific parameters.
We evaluate our method on a suite of fine-tuning tasks and architectures (ResNet, DenseNet, ViT) and show that it achieves state-of-the-art performance while being simple to implement.
arXiv Detail & Related papers (2022-03-30T23:16:07Z) - Multi-Task Learning with Sequence-Conditioned Transporter Networks [67.57293592529517]
We aim to solve multi-task learning through the lens of sequence-conditioning and weighted sampling.
We propose a new suite of benchmark aimed at compositional tasks, MultiRavens, which allows defining custom task combinations.
Second, we propose a vision-based end-to-end system architecture, Sequence-Conditioned Transporter Networks, which augments Goal-Conditioned Transporter Networks with sequence-conditioning and weighted sampling.
arXiv Detail & Related papers (2021-09-15T21:19:11Z) - Exploring Relational Context for Multi-Task Dense Prediction [76.86090370115]
We consider a multi-task environment for dense prediction tasks, represented by a common backbone and independent task-specific heads.
We explore various attention-based contexts, such as global and local, in the multi-task setting.
We propose an Adaptive Task-Relational Context module, which samples the pool of all available contexts for each task pair.
arXiv Detail & Related papers (2021-04-28T16:45:56Z) - Low Resource Multi-Task Sequence Tagging -- Revisiting Dynamic
Conditional Random Fields [67.51177964010967]
We compare different models for low resource multi-task sequence tagging that leverage dependencies between label sequences for different tasks.
We find that explicit modeling of inter-dependencies between task predictions outperforms single-task as well as standard multi-task models.
arXiv Detail & Related papers (2020-05-01T07:11:34Z) - Meta Reinforcement Learning with Autonomous Inference of Subtask
Dependencies [57.27944046925876]
We propose and address a novel few-shot RL problem, where a task is characterized by a subtask graph.
Instead of directly learning a meta-policy, we develop a Meta-learner with Subtask Graph Inference.
Our experiment results on two grid-world domains and StarCraft II environments show that the proposed method is able to accurately infer the latent task parameter.
arXiv Detail & Related papers (2020-01-01T17:34:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.