Multi-Task Learning via Co-Attentive Sharing for Pedestrian Attribute
Recognition
- URL: http://arxiv.org/abs/2004.03164v1
- Date: Tue, 7 Apr 2020 07:24:22 GMT
- Title: Multi-Task Learning via Co-Attentive Sharing for Pedestrian Attribute
Recognition
- Authors: Haitian Zeng, Haizhou Ai, Zijie Zhuang, Long Chen
- Abstract summary: Co-Attentive Sharing (CAS) module extracts discriminative channels and spatial regions for more effective feature sharing in multi-task learning.
Our module outperforms the conventional sharing units and achieves superior results compared to the state-of-the-art approaches using many metrics.
- Score: 8.883961218702824
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning to predict multiple attributes of a pedestrian is a multi-task
learning problem. To share feature representation between two individual task
networks, conventional methods like Cross-Stitch and Sluice network learn a
linear combination of features or feature subspaces. However, linear
combination rules out the complex interdependency between channels. Moreover,
spatial information exchanging is less-considered. In this paper, we propose a
novel Co-Attentive Sharing (CAS) module which extracts discriminative channels
and spatial regions for more effective feature sharing in multi-task learning.
The module consists of three branches, which leverage different channels for
between-task feature fusing, attention generation and task-specific feature
enhancing, respectively. Experiments on two pedestrian attribute recognition
datasets show that our module outperforms the conventional sharing units and
achieves superior results compared to the state-of-the-art approaches using
many metrics.
Related papers
- Multi-Task Learning for Fatigue Detection and Face Recognition of Drivers via Tree-Style Space-Channel Attention Fusion Network [7.6695642174485705]
In driving scenarios, automobile active safety systems are increasingly incorporating deep learning technology.
These systems typically need to handle multiple tasks simultaneously, such as detecting fatigue driving and recognizing the driver's identity.
We propose a novel tree-style multi-task modeling approach for multi-task learning, which rooted at a shared backbone.
arXiv Detail & Related papers (2024-05-13T15:34:20Z) - Sequential Cross Attention Based Multi-task Learning [22.430705836627148]
We propose a novel architecture that effectively transfers informative features by applying the attention mechanism to the multi-scale features of the tasks.
Our method achieves state-of-the-art performance on the NYUD-v2 and PASCAL-Context dataset.
arXiv Detail & Related papers (2022-09-06T14:17:33Z) - DenseMTL: Cross-task Attention Mechanism for Dense Multi-task Learning [18.745373058797714]
We propose a novel multi-task learning architecture that leverages pairwise cross-task exchange through correlation-guided attention and self-attention.
We conduct extensive experiments across three multi-task setups, showing the advantages of our approach compared to competitive baselines in both synthetic and real-world benchmarks.
arXiv Detail & Related papers (2022-06-17T17:59:45Z) - Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners [67.5865966762559]
We study whether sparsely activated Mixture-of-Experts (MoE) improve multi-task learning.
We devise task-aware gating functions to route examples from different tasks to specialized experts.
This results in a sparsely activated multi-task model with a large number of parameters, but with the same computational cost as that of a dense model.
arXiv Detail & Related papers (2022-04-16T00:56:12Z) - Channel Exchanging Networks for Multimodal and Multitask Dense Image
Prediction [125.18248926508045]
We propose Channel-Exchanging-Network (CEN) which is self-adaptive, parameter-free, and more importantly, applicable for both multimodal fusion and multitask learning.
CEN dynamically exchanges channels betweenworks of different modalities.
For the application of dense image prediction, the validity of CEN is tested by four different scenarios.
arXiv Detail & Related papers (2021-12-04T05:47:54Z) - Specificity-preserving RGB-D Saliency Detection [103.3722116992476]
We propose a specificity-preserving network (SP-Net) for RGB-D saliency detection.
Two modality-specific networks and a shared learning network are adopted to generate individual and shared saliency maps.
Experiments on six benchmark datasets demonstrate that our SP-Net outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2021-08-18T14:14:22Z) - Adversarial Semi-Supervised Multi-Domain Tracking [1.0152838128195465]
In visual tracking, the emerging features in shared layers of a multi-domain tracker are crucial for tracking in unseen videos.
We propose a semi-supervised learning scheme to separate domain-invariant and domain-specific features using adversarial learning.
We build a tracker that performs exceptionally on different types of videos.
arXiv Detail & Related papers (2020-09-30T12:47:28Z) - Multi-Task Learning with Deep Neural Networks: A Survey [0.0]
Multi-task learning (MTL) is a subfield of machine learning in which multiple tasks are simultaneously learned by a shared model.
We give an overview of multi-task learning methods for deep neural networks, with the aim of summarizing both the well-established and most recent directions within the field.
arXiv Detail & Related papers (2020-09-10T19:31:04Z) - Reparameterizing Convolutions for Incremental Multi-Task Learning
without Task Interference [75.95287293847697]
Two common challenges in developing multi-task models are often overlooked in literature.
First, enabling the model to be inherently incremental, continuously incorporating information from new tasks without forgetting the previously learned ones (incremental learning)
Second, eliminating adverse interactions amongst tasks, which has been shown to significantly degrade the single-task performance in a multi-task setup (task interference)
arXiv Detail & Related papers (2020-07-24T14:44:46Z) - Adversarial Continual Learning [99.56738010842301]
We propose a hybrid continual learning framework that learns a disjoint representation for task-invariant and task-specific features.
Our model combines architecture growth to prevent forgetting of task-specific skills and an experience replay approach to preserve shared skills.
arXiv Detail & Related papers (2020-03-21T02:08:17Z) - Unpaired Multi-modal Segmentation via Knowledge Distillation [77.39798870702174]
We propose a novel learning scheme for unpaired cross-modality image segmentation.
In our method, we heavily reuse network parameters, by sharing all convolutional kernels across CT and MRI.
We have extensively validated our approach on two multi-class segmentation problems.
arXiv Detail & Related papers (2020-01-06T20:03:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.