STAR-GNN: Spatial-Temporal Video Representation for Content-based
Retrieval
- URL: http://arxiv.org/abs/2208.06966v1
- Date: Mon, 15 Aug 2022 01:47:50 GMT
- Title: STAR-GNN: Spatial-Temporal Video Representation for Content-based
Retrieval
- Authors: Guoping Zhao, Bingqing Zhang, Mingyu Zhang, Yaxian Li, Jiajun Liu, and
Ji-Rong Wen
- Abstract summary: We propose a video feature representation learning framework called STAR-GNN.
It applies a pluggable graph neural network component on a multi-scale lattice feature graph.
We show that STAR-GNN effectively implements a dynamic attention mechanism on video frame sequences.
- Score: 39.50179338831056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a video feature representation learning framework called STAR-GNN,
which applies a pluggable graph neural network component on a multi-scale
lattice feature graph. The essence of STAR-GNN is to exploit both the temporal
dynamics and spatial contents as well as visual connections between regions at
different scales in the frames. It models a video with a lattice feature graph
in which the nodes represent regions of different granularity, with weighted
edges that represent the spatial and temporal links. The contextual nodes are
aggregated simultaneously by graph neural networks with parameters trained with
retrieval triplet loss. In the experiments, we show that STAR-GNN effectively
implements a dynamic attention mechanism on video frame sequences, resulting in
the emphasis for dynamic and semantically rich content in the video, and is
robust to noise and redundancies. Empirical results show that STAR-GNN achieves
state-of-the-art performance for Content-Based Video Retrieval.
Related papers
- Structure Enhanced Graph Neural Networks for Link Prediction [6.872826041648584]
We propose Structure Enhanced Graph neural network (SEG) for link prediction.
SEG incorporates surrounding topological information of target nodes into an ordinary GNN model.
Experiments on the OGB link prediction datasets demonstrate that SEG achieves state-of-the-art results.
arXiv Detail & Related papers (2022-01-14T03:49:30Z) - Graph Neural Networks with Feature and Structure Aware Random Walk [7.143879014059894]
We show that in typical heterphilous graphs, the edges may be directed, and whether to treat the edges as is or simply make them undirected greatly affects the performance of the GNN models.
We develop a model that adaptively learns the directionality of the graph, and exploits the underlying long-distance correlations between nodes.
arXiv Detail & Related papers (2021-11-19T08:54:21Z) - SEA: Graph Shell Attention in Graph Neural Networks [8.565134944225491]
A common issue in Graph Neural Networks (GNNs) is known as over-smoothing.
In our work, we relax the GNN architecture by means of implementing a routing. Specifically, the nodes' representations are routed to dedicated experts.
We call this procedure Graph Shell Attention (SEA), where experts process different subgraphs in a transformer-motivated fashion.
arXiv Detail & Related papers (2021-10-20T17:32:08Z) - Feature Correlation Aggregation: on the Path to Better Graph Neural
Networks [37.79964911718766]
Prior to the introduction of Graph Neural Networks (GNNs), modeling and analyzing irregular data, particularly graphs, was thought to be the Achilles' heel of deep learning.
This paper introduces a central node permutation variant function through a frustratingly simple and innocent-looking modification to the core operation of a GNN.
A tangible boost in performance of the model is observed where the model surpasses previous state-of-the-art results by a significant margin while employing fewer parameters.
arXiv Detail & Related papers (2021-09-20T05:04:26Z) - Node Similarity Preserving Graph Convolutional Networks [51.520749924844054]
Graph Neural Networks (GNNs) explore the graph structure and node features by aggregating and transforming information within node neighborhoods.
We propose SimP-GCN that can effectively and efficiently preserve node similarity while exploiting graph structure.
We validate the effectiveness of SimP-GCN on seven benchmark datasets including three assortative and four disassorative graphs.
arXiv Detail & Related papers (2020-11-19T04:18:01Z) - Video-based Facial Expression Recognition using Graph Convolutional
Networks [57.980827038988735]
We introduce a Graph Convolutional Network (GCN) layer into a common CNN-RNN based model for video-based facial expression recognition.
We evaluate our method on three widely-used datasets, CK+, Oulu-CASIA and MMI, and also one challenging wild dataset AFEW8.0.
arXiv Detail & Related papers (2020-10-26T07:31:51Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Structural Temporal Graph Neural Networks for Anomaly Detection in
Dynamic Graphs [54.13919050090926]
We propose an end-to-end structural temporal Graph Neural Network model for detecting anomalous edges in dynamic graphs.
In particular, we first extract the $h$-hop enclosing subgraph centered on the target edge and propose the node labeling function to identify the role of each node in the subgraph.
Based on the extracted features, we utilize Gated recurrent units (GRUs) to capture the temporal information for anomaly detection.
arXiv Detail & Related papers (2020-05-15T09:17:08Z) - Zero-Shot Video Object Segmentation via Attentive Graph Neural Networks [150.5425122989146]
This work proposes a novel attentive graph neural network (AGNN) for zero-shot video object segmentation (ZVOS)
AGNN builds a fully connected graph to efficiently represent frames as nodes, and relations between arbitrary frame pairs as edges.
Experimental results on three video segmentation datasets show that AGNN sets a new state-of-the-art in each case.
arXiv Detail & Related papers (2020-01-19T10:45:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.