All About Knowledge Graphs for Actions
- URL: http://arxiv.org/abs/2008.12432v1
- Date: Fri, 28 Aug 2020 01:44:01 GMT
- Title: All About Knowledge Graphs for Actions
- Authors: Pallabi Ghosh, Nirat Saini, Larry S. Davis, Abhinav Shrivastava
- Abstract summary: We propose a better understanding of knowledge graphs (KGs) that can be utilized for zero-shot and few-shot action recognition.
We study three different construction mechanisms for KGs: action embeddings, action-object embeddings, visual embeddings.
We present extensive analysis of the impact of different KGs on different experimental setups.
- Score: 82.39684757372075
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current action recognition systems require large amounts of training data for
recognizing an action. Recent works have explored the paradigm of zero-shot and
few-shot learning to learn classifiers for unseen categories or categories with
few labels. Following similar paradigms in object recognition, these approaches
utilize external sources of knowledge (eg. knowledge graphs from language
domains). However, unlike objects, it is unclear what is the best knowledge
representation for actions. In this paper, we intend to gain a better
understanding of knowledge graphs (KGs) that can be utilized for zero-shot and
few-shot action recognition. In particular, we study three different
construction mechanisms for KGs: action embeddings, action-object embeddings,
visual embeddings. We present extensive analysis of the impact of different KGs
in different experimental setups. Finally, to enable a systematic study of
zero-shot and few-shot approaches, we propose an improved evaluation paradigm
based on UCF101, HMDB51, and Charades datasets for knowledge transfer from
models trained on Kinetics.
Related papers
- Knowledge-Aware Prompt Tuning for Generalizable Vision-Language Models [64.24227572048075]
We propose a Knowledge-Aware Prompt Tuning (KAPT) framework for vision-language models.
Our approach takes inspiration from human intelligence in which external knowledge is usually incorporated into recognizing novel categories of objects.
arXiv Detail & Related papers (2023-08-22T04:24:45Z) - Sample-Efficient Learning of Novel Visual Concepts [7.398195748292981]
State-of-the-art deep learning models struggle to recognize novel objects in a few-shot setting.
We show that incorporating a symbolic knowledge graph into a state-of-the-art recognition model enables a new approach for effective few-shot classification.
arXiv Detail & Related papers (2023-06-15T20:24:30Z) - Recognizing Unseen Objects via Multimodal Intensive Knowledge Graph
Propagation [68.13453771001522]
We propose a multimodal intensive ZSL framework that matches regions of images with corresponding semantic embeddings.
We conduct extensive experiments and evaluate our model on large-scale real-world data.
arXiv Detail & Related papers (2023-06-14T13:07:48Z) - Hint-dynamic Knowledge Distillation [30.40008256306688]
Hint-dynamic Knowledge Distillation, dubbed HKD, excavates the knowledge from the teacher's hints in a dynamic scheme.
A meta-weight network is introduced to generate the instance-wise weight coefficients about knowledge hints.
Experiments on standard benchmarks of CIFAR-100 and Tiny-ImageNet manifest that the proposed HKD well boost the effect of knowledge distillation tasks.
arXiv Detail & Related papers (2022-11-30T15:03:53Z) - A Survey on Visual Transfer Learning using Knowledge Graphs [0.8701566919381223]
This survey focuses on visual transfer learning approaches using knowledge graphs (KGs)
KGs can represent auxiliary knowledge either in an underlying graph-structured schema or in a vector-based knowledge graph embedding.
We provide a broad overview of knowledge graph embedding methods and describe several joint training objectives suitable to combine them with high dimensional visual embeddings.
arXiv Detail & Related papers (2022-01-27T20:19:55Z) - Ontology-enhanced Prompt-tuning for Few-shot Learning [41.51144427728086]
Few-shot Learning is aimed to make predictions based on a limited number of samples.
Structured data such as knowledge graphs and ontology libraries has been leveraged to benefit the few-shot setting in various tasks.
arXiv Detail & Related papers (2022-01-27T05:41:36Z) - Reasoning over Vision and Language: Exploring the Benefits of
Supplemental Knowledge [59.87823082513752]
This paper investigates the injection of knowledge from general-purpose knowledge bases (KBs) into vision-and-language transformers.
We empirically study the relevance of various KBs to multiple tasks and benchmarks.
The technique is model-agnostic and can expand the applicability of any vision-and-language transformer with minimal computational overhead.
arXiv Detail & Related papers (2021-01-15T08:37:55Z) - Memory-augmented Dense Predictive Coding for Video Representation
Learning [103.69904379356413]
We propose a new architecture and learning framework Memory-augmented Predictive Coding (MemDPC) for the task.
We investigate visual-only self-supervised video representation learning from RGB frames, or from unsupervised optical flow, or both.
In all cases, we demonstrate state-of-the-art or comparable performance over other approaches with orders of magnitude fewer training data.
arXiv Detail & Related papers (2020-08-03T17:57:01Z) - KACC: A Multi-task Benchmark for Knowledge Abstraction, Concretization
and Completion [99.47414073164656]
A comprehensive knowledge graph (KG) contains an instance-level entity graph and an ontology-level concept graph.
The two-view KG provides a testbed for models to "simulate" human's abilities on knowledge abstraction, concretization, and completion.
We propose a unified KG benchmark by improving existing benchmarks in terms of dataset scale, task coverage, and difficulty.
arXiv Detail & Related papers (2020-04-28T16:21:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.