Differentiable Reasoning over Long Stories -- Assessing Systematic
Generalisation in Neural Models
- URL: http://arxiv.org/abs/2203.10620v1
- Date: Sun, 20 Mar 2022 18:34:42 GMT
- Title: Differentiable Reasoning over Long Stories -- Assessing Systematic
Generalisation in Neural Models
- Authors: Wanshui Li, Pasquale Minervini
- Abstract summary: We consider two classes of neural models: "E-GNN", the graph-based models that can process graph-structured data and consider the edge attributes simultaneously; and "L-Graph", the sequence-based models which can process linearized version of the graphs.
We found that the modified recurrent neural network yield surprisingly accurate results across every systematic generalisation tasks which outperform the graph neural network.
- Score: 12.479512369785082
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contemporary neural networks have achieved a series of developments and
successes in many aspects; however, when exposed to data outside the training
distribution, they may fail to predict correct answers. In this work, we were
concerned about this generalisation issue and thus analysed a broad set of
models systematically and robustly over long stories. Related experiments were
conducted based on the CLUTRR, which is a diagnostic benchmark suite that can
analyse generalisation of natural language understanding (NLU) systems by
training over small story graphs and testing on larger ones. In order to handle
the multi-relational story graph, we consider two classes of neural models:
"E-GNN", the graph-based models that can process graph-structured data and
consider the edge attributes simultaneously; and "L-Graph", the sequence-based
models which can process linearized version of the graphs. We performed an
extensive empirical evaluation, and we found that the modified recurrent neural
network yield surprisingly accurate results across every systematic
generalisation tasks which outperform the modified graph neural network, while
the latter produced more robust models.
Related papers
- Do graph neural network states contain graph properties? [5.222978725954348]
We present a model explainability pipeline for Graph Neural Networks (GNNs) employing diagnostic classifiers.
This pipeline aims to probe and interpret the learned representations in GNNs across various architectures and datasets.
arXiv Detail & Related papers (2024-11-04T15:26:07Z) - Self-Supervised Graph Neural Networks for Enhanced Feature Extraction in Heterogeneous Information Networks [16.12856816023414]
This paper explores the applications and challenges of graph neural networks (GNNs) in processing complex graph data brought about by the rapid development of the Internet.
By introducing a self-supervisory mechanism, it is expected to improve the adaptability of existing models to the diversity and complexity of graph data.
arXiv Detail & Related papers (2024-10-23T07:14:37Z) - Generalization of Graph Neural Networks is Robust to Model Mismatch [84.01980526069075]
Graph neural networks (GNNs) have demonstrated their effectiveness in various tasks supported by their generalization capabilities.
In this paper, we examine GNNs that operate on geometric graphs generated from manifold models.
Our analysis reveals the robustness of the GNN generalization in the presence of such model mismatch.
arXiv Detail & Related papers (2024-08-25T16:00:44Z) - Challenging the Myth of Graph Collaborative Filtering: a Reasoned and Reproducibility-driven Analysis [50.972595036856035]
We present a code that successfully replicates results from six popular and recent graph recommendation models.
We compare these graph models with traditional collaborative filtering models that historically performed well in offline evaluations.
By investigating the information flow from users' neighborhoods, we aim to identify which models are influenced by intrinsic features in the dataset structure.
arXiv Detail & Related papers (2023-08-01T09:31:44Z) - GraphGLOW: Universal and Generalizable Structure Learning for Graph
Neural Networks [72.01829954658889]
This paper introduces the mathematical definition of this novel problem setting.
We devise a general framework that coordinates a single graph-shared structure learner and multiple graph-specific GNNs.
The well-trained structure learner can directly produce adaptive structures for unseen target graphs without any fine-tuning.
arXiv Detail & Related papers (2023-06-20T03:33:22Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - GDBN: a Graph Neural Network Approach to Dynamic Bayesian Network [7.876789380671075]
We propose a graph neural network approach with score-based method aiming at learning a sparse DAG.
We demonstrate methods with graph neural network significantly outperformed other state-of-the-art methods with dynamic bayesian networking inference.
arXiv Detail & Related papers (2023-01-28T02:49:13Z) - Learning the Evolutionary and Multi-scale Graph Structure for
Multivariate Time Series Forecasting [50.901984244738806]
We show how to model the evolutionary and multi-scale interactions of time series.
In particular, we first provide a hierarchical graph structure cooperated with the dilated convolution to capture the scale-specific correlations.
A unified neural network is provided to integrate the components above to get the final prediction.
arXiv Detail & Related papers (2022-06-28T08:11:12Z) - Hyperbolic Graph Neural Networks: A Review of Methods and Applications [55.5502008501764]
Graph neural networks generalize conventional neural networks to graph-structured data.
The performance of Euclidean models in graph-related learning is still bounded and limited by the representation ability of Euclidean geometry.
Recently, hyperbolic space has gained increasing popularity in processing graph data with tree-like structure and power-law distribution.
arXiv Detail & Related papers (2022-02-28T15:08:48Z) - Crime Prediction with Graph Neural Networks and Multivariate Normal
Distributions [18.640610803366876]
We tackle the sparsity problem in high resolution by leveraging the flexible structure of graph convolutional networks (GCNs)
We build our model with Graph Convolutional Gated Recurrent Units (Graph-ConvGRU) to learn spatial, temporal, and categorical relations.
We show that our model is not only generative but also precise.
arXiv Detail & Related papers (2021-11-29T17:37:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.