Knowledge as Priors: Cross-Modal Knowledge Generalization for Datasets
without Superior Knowledge
- URL: http://arxiv.org/abs/2004.00176v1
- Date: Wed, 1 Apr 2020 00:28:15 GMT
- Title: Knowledge as Priors: Cross-Modal Knowledge Generalization for Datasets
without Superior Knowledge
- Authors: Long Zhao, Xi Peng, Yuxiao Chen, Mubbasir Kapadia, Dimitris N. Metaxas
- Abstract summary: Cross-modal knowledge distillation deals with transferring knowledge from a model trained with superior modalities to another model trained with weak modalities.
We propose a novel scheme to train the Student in a Target dataset where the Teacher is unavailable.
- Score: 55.32035138692167
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cross-modal knowledge distillation deals with transferring knowledge from a
model trained with superior modalities (Teacher) to another model trained with
weak modalities (Student). Existing approaches require paired training examples
exist in both modalities. However, accessing the data from superior modalities
may not always be feasible. For example, in the case of 3D hand pose
estimation, depth maps, point clouds, or stereo images usually capture better
hand structures than RGB images, but most of them are expensive to be
collected. In this paper, we propose a novel scheme to train the Student in a
Target dataset where the Teacher is unavailable. Our key idea is to generalize
the distilled cross-modal knowledge learned from a Source dataset, which
contains paired examples from both modalities, to the Target dataset by
modeling knowledge as priors on parameters of the Student. We name our method
"Cross-Modal Knowledge Generalization" and demonstrate that our scheme results
in competitive performance for 3D hand pose estimation on standard benchmark
datasets.
Related papers
- Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - Towards a Generalist and Blind RGB-X Tracker [91.36268768952755]
We develop a single model tracker that can remain blind to any modality X during inference time.
Our training process is extremely simple, integrating multi-label classification loss with a routing function.
Our generalist and blind tracker can achieve competitive performance compared to well-established modal-specific models.
arXiv Detail & Related papers (2024-05-28T03:00:58Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - Cross-Modal Fine-Tuning: Align then Refine [83.37294254884446]
ORCA is a cross-modal fine-tuning framework that extends the applicability of a single large-scale pretrained model to diverse modalities.
We show that ORCA obtains state-of-the-art results on 3 benchmarks containing over 60 datasets from 12 modalities.
arXiv Detail & Related papers (2023-02-11T16:32:28Z) - CMD: Self-supervised 3D Action Representation Learning with Cross-modal
Mutual Distillation [130.08432609780374]
In 3D action recognition, there exists rich complementary information between skeleton modalities.
We propose a new Cross-modal Mutual Distillation (CMD) framework with the following designs.
Our approach outperforms existing self-supervised methods and sets a series of new records.
arXiv Detail & Related papers (2022-08-26T06:06:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.