Beyond the Gates of Euclidean Space: Temporal-Discrimination-Fusions and
Attention-based Graph Neural Network for Human Activity Recognition
- URL: http://arxiv.org/abs/2206.04855v1
- Date: Fri, 10 Jun 2022 03:04:23 GMT
- Title: Beyond the Gates of Euclidean Space: Temporal-Discrimination-Fusions and
Attention-based Graph Neural Network for Human Activity Recognition
- Authors: Nafees Ahmad, Savio Ho-Chit Chow, Ho-fung Leung
- Abstract summary: Human activity recognition (HAR) through wearable devices has received much interest due to its numerous applications in fitness tracking, wellness screening, and supported living.
Traditional deep learning (DL) has set a state of the art performance for HAR domain.
We propose an approach based on Graph Neural Networks (GNNs) for structuring the input representation and exploiting the relations among the samples.
- Score: 5.600003119721707
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human activity recognition (HAR) through wearable devices has received much
interest due to its numerous applications in fitness tracking, wellness
screening, and supported living. As a result, we have seen a great deal of work
in this field. Traditional deep learning (DL) has set a state of the art
performance for HAR domain. However, it ignores the data's structure and the
association between consecutive time stamps. To address this constraint, we
offer an approach based on Graph Neural Networks (GNNs) for structuring the
input representation and exploiting the relations among the samples. However,
even when using a simple graph convolution network to eliminate this shortage,
there are still several limiting factors, such as inter-class activities
issues, skewed class distribution, and a lack of consideration for sensor data
priority, all of which harm the HAR model's performance. To improve the current
HAR model's performance, we investigate novel possibilities within the
framework of graph structure to achieve highly discriminated and rich activity
features. We propose a model for (1) time-series-graph module that converts raw
data from HAR dataset into graphs; (2) Graph Convolutional Neural Networks
(GCNs) to discover local dependencies and correlations between neighboring
nodes; and (3) self-attention GNN encoder to identify sensors interactions and
data priorities. To the best of our knowledge, this is the first work for HAR,
which introduces a GNN-based approach that incorporates both the GCN and the
attention mechanism. By employing a uniform evaluation method, our framework
significantly improves the performance on hospital patient's activities dataset
comparatively considered other state of the art baseline methods.
Related papers
- Graph as a feature: improving node classification with non-neural graph-aware logistic regression [2.952177779219163]
Graph-aware Logistic Regression (GLR) is a non-neural model designed for node classification tasks.
Unlike traditional graph algorithms that use only a fraction of the information accessible to GNNs, our proposed model simultaneously leverages both node features and the relationships between entities.
arXiv Detail & Related papers (2024-11-19T08:32:14Z) - Heterogeneous Hyper-Graph Neural Networks for Context-aware Human Activity Recognition [2.8132886759540146]
We argue that context-aware activity visit patterns in realistic in-the-wild data can be considered as a general graph representation learning task.
We propose a novel Heterogeneous HyperGraph Neural Network architecture for Context-aware Human Activity Recognition.
arXiv Detail & Related papers (2024-09-26T02:44:37Z) - Label Deconvolution for Node Representation Learning on Large-scale
Attributed Graphs against Learning Bias [75.44877675117749]
We propose an efficient label regularization technique, namely Label Deconvolution (LD), to alleviate the learning bias by a novel and highly scalable approximation to the inverse mapping of GNNs.
Experiments demonstrate LD significantly outperforms state-of-the-art methods on Open Graph datasets Benchmark.
arXiv Detail & Related papers (2023-09-26T13:09:43Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - Task-Agnostic Graph Neural Network Evaluation via Adversarial
Collaboration [11.709808788756966]
GraphAC is a principled, task-agnostic, and stable framework for evaluating Graph Neural Network (GNN) research for molecular representation learning.
We introduce a novel objective function: the Competitive Barlow Twins, that allow two GNNs to jointly update themselves from direct competitions against each other.
arXiv Detail & Related papers (2023-01-27T03:33:11Z) - Dynamic Graph Message Passing Networks for Visual Recognition [112.49513303433606]
Modelling long-range dependencies is critical for scene understanding tasks in computer vision.
A fully-connected graph is beneficial for such modelling, but its computational overhead is prohibitive.
We propose a dynamic graph message passing network, that significantly reduces the computational complexity.
arXiv Detail & Related papers (2022-09-20T14:41:37Z) - Affinity-Aware Graph Networks [9.888383815189176]
Graph Neural Networks (GNNs) have emerged as a powerful technique for learning on relational data.
We explore the use of affinity measures as features in graph neural networks.
We propose message passing networks based on these features and evaluate their performance on a variety of node and graph property prediction tasks.
arXiv Detail & Related papers (2022-06-23T18:51:35Z) - Automatic Relation-aware Graph Network Proliferation [182.30735195376792]
We propose Automatic Relation-aware Graph Network Proliferation (ARGNP) for efficiently searching GNNs.
These operations can extract hierarchical node/relational information and provide anisotropic guidance for message passing on a graph.
Experiments on six datasets for four graph learning tasks demonstrate that GNNs produced by our method are superior to the current state-of-the-art hand-crafted and search-based GNNs.
arXiv Detail & Related papers (2022-05-31T10:38:04Z) - Discovering the Representation Bottleneck of Graph Neural Networks from
Multi-order Interactions [51.597480162777074]
Graph neural networks (GNNs) rely on the message passing paradigm to propagate node features and build interactions.
Recent works point out that different graph learning tasks require different ranges of interactions between nodes.
We study two common graph construction methods in scientific domains, i.e., emphK-nearest neighbor (KNN) graphs and emphfully-connected (FC) graphs.
arXiv Detail & Related papers (2022-05-15T11:38:14Z) - Graph Neural Networks for Inconsistent Cluster Detection in Incremental
Entity Resolution [3.4806267677524896]
In mature data repositories, the relationships may be mostly correct but require incremental improvements owing to errors in the original data or in the entity resolution system.
This paper proposes a novel method for identifying inconsistent clusters (IC), existing groups of related products that do not belong together.
We demonstrate that existing Message Passing neural networks perform well at this task, exceeding traditional graph processing techniques.
arXiv Detail & Related papers (2021-05-12T20:39:22Z) - Graph Representation Learning via Graphical Mutual Information
Maximization [86.32278001019854]
We propose a novel concept, Graphical Mutual Information (GMI), to measure the correlation between input graphs and high-level hidden representations.
We develop an unsupervised learning model trained by maximizing GMI between the input and output of a graph neural encoder.
arXiv Detail & Related papers (2020-02-04T08:33:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.