A Dynamic Graph CNN with Cross-Representation Distillation for
Event-Based Recognition
- URL: http://arxiv.org/abs/2302.04177v2
- Date: Sun, 16 Apr 2023 08:14:29 GMT
- Title: A Dynamic Graph CNN with Cross-Representation Distillation for
Event-Based Recognition
- Authors: Yongjian Deng, Hao Chen, Bochen Xie, Hai Liu, Youfu Li
- Abstract summary: We present a new event-based graph learning framework called graph cross-representation distillation (CRD)
CRD provides additional supervision and prior knowledge for the event graph.
Our model and learning framework are effective and generalize well across multiple vision tasks.
- Score: 21.225945234873745
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advances in event-based research prioritize sparsity and temporal
precision. Approaches using dense frame-based representations processed via
well-pretrained CNNs are being replaced by the use of sparse point-based
representations learned through graph CNNs (GCN). Yet, the efficacy of these
graph methods is far behind their frame-based counterparts with two
limitations. ($i$) Biased graph construction without carefully integrating
variant attributes ($i.e.$, semantics, spatial and temporal cues) for each
vertex, leading to imprecise graph representation. ($ii$) Deficient learning
because of the lack of well-pretrained models available. Here we solve the
first problem by proposing a new event-based GCN (EDGCN), with a dynamic
aggregation module to integrate all attributes of vertices adaptively. To
address the second problem, we introduce a novel learning framework called
cross-representation distillation (CRD), which leverages the dense
representation of events as a cross-representation auxiliary to provide
additional supervision and prior knowledge for the event graph. This
frame-to-graph distillation allows us to benefit from the large-scale priors
provided by CNNs while still retaining the advantages of graph-based models.
Extensive experiments show our model and learning framework are effective and
generalize well across multiple vision tasks.
Related papers
- A Pure Transformer Pretraining Framework on Text-attributed Graphs [50.833130854272774]
We introduce a feature-centric pretraining perspective by treating graph structure as a prior.
Our framework, Graph Sequence Pretraining with Transformer (GSPT), samples node contexts through random walks.
GSPT can be easily adapted to both node classification and link prediction, demonstrating promising empirical success on various datasets.
arXiv Detail & Related papers (2024-06-19T22:30:08Z) - Input Snapshots Fusion for Scalable Discrete Dynamic Graph Nerual Networks [27.616083395612595]
We introduce an Input bf Snapshots bf Fusion based bf Dynamic bf Graph Neural Network (SFDyG)
By eliminating the partitioning of snapshots within the input window, we obtain a multi-graph (more than one edge between two nodes)
We propose a scalable three-step mini-batch training method and demonstrate its equivalence to full-batch training counterpart.
arXiv Detail & Related papers (2024-05-11T10:05:55Z) - SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - Causal Incremental Graph Convolution for Recommender System Retraining [89.25922726558875]
Real-world recommender system needs to be regularly retrained to keep with the new data.
In this work, we consider how to efficiently retrain graph convolution network (GCN) based recommender models.
arXiv Detail & Related papers (2021-08-16T04:20:09Z) - TrackMPNN: A Message Passing Graph Neural Architecture for Multi-Object
Tracking [8.791710193028903]
This study follows many previous approaches to multi-object tracking (MOT) that model the problem using graph-based data structures.
We create a framework based on dynamic undirected graphs that represent the data association problem over multiple timesteps.
We also provide solutions and propositions for the computational problems that need to be addressed to create a memory-efficient, real-time, online algorithm.
arXiv Detail & Related papers (2021-01-11T21:52:25Z) - Sub-graph Contrast for Scalable Self-Supervised Graph Representation
Learning [21.0019144298605]
Existing graph neural networks fed with the complete graph data are not scalable due to limited computation and memory costs.
textscSubg-Con is proposed by utilizing the strong correlation between central nodes and their sampled subgraphs to capture regional structure information.
Compared with existing graph representation learning approaches, textscSubg-Con has prominent performance advantages in weaker supervision requirements, model learning scalability, and parallelization.
arXiv Detail & Related papers (2020-09-22T01:58:19Z) - GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training [62.73470368851127]
Graph representation learning has emerged as a powerful technique for addressing real-world problems.
We design Graph Contrastive Coding -- a self-supervised graph neural network pre-training framework.
We conduct experiments on three graph learning tasks and ten graph datasets.
arXiv Detail & Related papers (2020-06-17T16:18:35Z) - Graph Ordering: Towards the Optimal by Learning [69.72656588714155]
Graph representation learning has achieved a remarkable success in many graph-based applications, such as node classification, prediction, and community detection.
However, for some kind of graph applications, such as graph compression and edge partition, it is very hard to reduce them to some graph representation learning tasks.
In this paper, we propose to attack the graph ordering problem behind such applications by a novel learning approach.
arXiv Detail & Related papers (2020-01-18T09:14:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.