Team PyKale (xy9) Submission to the EPIC-Kitchens 2021 Unsupervised
Domain Adaptation Challenge for Action Recognition
- URL: http://arxiv.org/abs/2106.12023v1
- Date: Tue, 22 Jun 2021 19:17:03 GMT
- Title: Team PyKale (xy9) Submission to the EPIC-Kitchens 2021 Unsupervised
Domain Adaptation Challenge for Action Recognition
- Authors: Xianyuan Liu, Raivo Koot, Shuo Zhou, Tao Lei, Haiping Lu
- Abstract summary: This report describes the technical details of our submission to the EPIC-Kitchens 2021 Unsupervised Domain Adaptation Challenge for Action Recognition.
The EPIC-Kitchens dataset is more difficult than other video domain adaptation datasets due to multi-tasks with more modalities.
Under the team name xy9, our submission achieved 5th place in terms of top-1 accuracy for verb class and all top-5 accuracies.
- Score: 12.905251261775405
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This report describes the technical details of our submission to the
EPIC-Kitchens 2021 Unsupervised Domain Adaptation Challenge for Action
Recognition. The EPIC-Kitchens dataset is more difficult than other video
domain adaptation datasets due to multi-tasks with more modalities. Firstly, to
participate in the challenge, we employ a transformer to capture the spatial
information from each modality. Secondly, we employ a temporal attention module
to model temporal-wise inter-dependency. Thirdly, we employ the adversarial
domain adaptation network to learn the general features between labeled source
and unlabeled target domain. Finally, we incorporate multiple modalities to
improve the performance by a three-stream network with late fusion. Our network
achieves the comparable performance with the state-of-the-art baseline T$A^3$N
and outperforms the baseline on top-1 accuracy for verb class and top-5
accuracies for all three tasks which are verb, noun and action. Under the team
name xy9, our submission achieved 5th place in terms of top-1 accuracy for verb
class and all top-5 accuracies.
Related papers
- Point-In-Context: Understanding Point Cloud via In-Context Learning [67.20277182808992]
We introduce Point-In-Context (PIC), a novel framework for 3D point cloud understanding via in-context learning.
We address the technical challenge of effectively extending masked point modeling to 3D point clouds by introducing a Joint Sampling module.
We propose two novel training strategies, In-Context Labeling and In-Context Enhancing, forming an extended version of PIC named Point-In-Context-Segmenter (PIC-S)
arXiv Detail & Related papers (2024-04-18T17:32:32Z) - POND: Multi-Source Time Series Domain Adaptation with Information-Aware Prompt Tuning [40.197245493051526]
Time series domain adaptation stands as a pivotal and intricate challenge with diverse applications.
We introduce PrOmpt-based domaiN Discrimination (POND), the first framework to utilize prompts for time series domain adaptation.
Our proposed POND model outperforms all state-of-the-art comparison methods by up to $66%$ on the F1-score.
arXiv Detail & Related papers (2023-12-19T15:57:37Z) - EPIC-KITCHENS-100 Unsupervised Domain Adaptation Challenge: Mixed
Sequences Prediction [16.92053939360415]
This report presents the technical details of our approach for the EPIC-Kitchens-100 Unsupervised Domain Adaptation (UDA) Challenge in Action Recognition.
Our approach is based on the idea that the order in which actions are performed is similar between the source and target domains.
We generate a modified sequence by randomly combining actions from the source and target domains.
arXiv Detail & Related papers (2023-07-24T14:35:46Z) - PoliTO-IIT-CINI Submission to the EPIC-KITCHENS-100 Unsupervised Domain
Adaptation Challenge for Action Recognition [16.496889090237232]
This report describes the technical details of our submission to the EPIC-Kitchens-100 Unsupervised Domain Adaptation Challenge in Action Recognition.
We first exploited a recent Domain Generalization technique, called Relative Norm Alignment (RNA)
Secondly, we extended this approach to work on unlabelled target data, enabling a simpler adaptation of the model to the target distribution in an unsupervised fashion.
arXiv Detail & Related papers (2022-09-09T21:03:11Z) - PoliTO-IIT Submission to the EPIC-KITCHENS-100 Unsupervised Domain
Adaptation Challenge for Action Recognition [15.545769463854915]
This report describes our submission to the EPIC-Kitchens-100 Unsupervised Domain Adaptation (UDA) Challenge in Action Recognition.
We first exploited a recent Domain Generalization (DG) technique, called Relative Norm Alignment (RNA)
In a second phase, we extended the approach to work on unlabelled target data, allowing the model to adapt to the target distribution in an unsupervised fashion.
Our submission (entry 'plnet') is visible on the leaderboard and it achieved the 1st position for'verb', and the 3rd position for both 'noun' and 'action'
arXiv Detail & Related papers (2021-07-01T10:02:44Z) - Curriculum Graph Co-Teaching for Multi-Target Domain Adaptation [78.28390172958643]
We identify two key aspects that can help to alleviate multiple domain-shifts in the multi-target domain adaptation (MTDA)
We propose Curriculum Graph Co-Teaching (CGCT) that uses a dual classifier head, with one of them being a graph convolutional network (GCN) which aggregates features from similar samples across the domains.
When the domain labels are available, we propose Domain-aware Curriculum Learning (DCL), a sequential adaptation strategy that first adapts on the easier target domains, followed by the harder ones.
arXiv Detail & Related papers (2021-04-01T23:41:41Z) - Learning Invariant Representations across Domains and Tasks [81.30046935430791]
We propose a novel Task Adaptation Network (TAN) to solve this unsupervised task transfer problem.
In addition to learning transferable features via domain-adversarial training, we propose a novel task semantic adaptor that uses the learning-to-learn strategy to adapt the task semantics.
TAN significantly increases the recall and F1 score by 5.0% and 7.8% compared to recently strong baselines.
arXiv Detail & Related papers (2021-03-03T11:18:43Z) - Cross-modal Learning for Domain Adaptation in 3D Semantic Segmentation [11.895722159139108]
Domain adaptation is an important task to enable learning when labels are scarce.
We propose cross-modal learning, where we enforce consistency between the predictions of two modalities via mutual mimicking.
We constrain our network to make correct predictions on labeled data and consistent predictions across modalities on unlabeled target-domain data.
arXiv Detail & Related papers (2021-01-18T18:59:21Z) - Deep Co-Training with Task Decomposition for Semi-Supervised Domain
Adaptation [80.55236691733506]
Semi-supervised domain adaptation (SSDA) aims to adapt models trained from a labeled source domain to a different but related target domain.
We propose to explicitly decompose the SSDA task into two sub-tasks: a semi-supervised learning (SSL) task in the target domain and an unsupervised domain adaptation (UDA) task across domains.
arXiv Detail & Related papers (2020-07-24T17:57:54Z) - Structured Domain Adaptation with Online Relation Regularization for
Unsupervised Person Re-ID [62.90727103061876]
Unsupervised domain adaptation (UDA) aims at adapting the model trained on a labeled source-domain dataset to an unlabeled target-domain dataset.
We propose an end-to-end structured domain adaptation framework with an online relation-consistency regularization term.
Our proposed framework is shown to achieve state-of-the-art performance on multiple UDA tasks of person re-ID.
arXiv Detail & Related papers (2020-03-14T14:45:18Z) - Continuous Domain Adaptation with Variational Domain-Agnostic Feature
Replay [78.7472257594881]
Learning in non-stationary environments is one of the biggest challenges in machine learning.
Non-stationarity can be caused by either task drift, or the domain drift.
We propose variational domain-agnostic feature replay, an approach that is composed of three components.
arXiv Detail & Related papers (2020-03-09T19:50:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.