Graph-based Neural Modules to Inspect Attention-based Architectures: A
Position Paper
- URL: http://arxiv.org/abs/2210.07117v1
- Date: Thu, 13 Oct 2022 15:52:12 GMT
- Title: Graph-based Neural Modules to Inspect Attention-based Architectures: A
Position Paper
- Authors: Breno W. Carvalho, Artur D'Avilla Garcez, Luis C. Lamb
- Abstract summary: encoder-decoder models offer an exciting opportunity for visualization and editing by humans of the knowledge implicitly represented in model weights.
In this work, we explore ways to create an abstraction for segments of the network as a two-way graph-based representation.
Such two-way graph representation enables new neuro-symbolic systems by leveraging the pattern recognition capabilities of the encoder-decoder along with symbolic reasoning carried out on the graphs.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Encoder-decoder architectures are prominent building blocks of
state-of-the-art solutions for tasks across multiple fields where deep learning
(DL) or foundation models play a key role. Although there is a growing
community working on the provision of interpretation for DL models as well as
considerable work in the neuro-symbolic community seeking to integrate symbolic
representations and DL, many open questions remain around the need for better
tools for visualization of the inner workings of DL architectures. In
particular, encoder-decoder models offer an exciting opportunity for
visualization and editing by humans of the knowledge implicitly represented in
model weights. In this work, we explore ways to create an abstraction for
segments of the network as a two-way graph-based representation. Changes to
this graph structure should be reflected directly in the underlying tensor
representations. Such two-way graph representation enables new neuro-symbolic
systems by leveraging the pattern recognition capabilities of the
encoder-decoder along with symbolic reasoning carried out on the graphs. The
approach is expected to produce new ways of interacting with DL models but also
to improve performance as a result of the combination of learning and reasoning
capabilities.
Related papers
- Towards Graph Foundation Models: A Survey and Beyond [66.37994863159861]
Foundation models have emerged as critical components in a variety of artificial intelligence applications.
The capabilities of foundation models to generalize and adapt motivate graph machine learning researchers to discuss the potential of developing a new graph learning paradigm.
This article introduces the concept of Graph Foundation Models (GFMs), and offers an exhaustive explanation of their key characteristics and underlying technologies.
arXiv Detail & Related papers (2023-10-18T09:31:21Z) - Serving Deep Learning Model in Relational Databases [70.53282490832189]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-art DL-centric architecture offloads DL computations to dedicated DL frameworks.
The potential UDF-centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the relational database management system (RDBMS)
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - Weisfeiler and Leman Go Relational [4.29881872550313]
We investigate the limitations in the expressive power of the well-known GCN and Composition GCN architectures.
We introduce the $k$-RN architecture that provably overcomes the limitations of the above two architectures.
arXiv Detail & Related papers (2022-11-30T15:56:46Z) - KGLM: Integrating Knowledge Graph Structure in Language Models for Link
Prediction [0.0]
We introduce a new entity/relation embedding layer that learns to differentiate distinctive entity and relation types.
We show that further pre-training the language models with this additional embedding layer using the triples extracted from the knowledge graph, followed by the standard fine-tuning phase sets a new state-of-the-art performance for the link prediction task on the benchmark datasets.
arXiv Detail & Related papers (2022-11-04T20:38:12Z) - Motif-based Graph Representation Learning with Application to Chemical
Molecules [11.257235936629689]
Existing graph neural networks offer limited ability to capture complex interactions within local structural contexts.
We propose a new motif-based graph representation learning technique to better utilize local structural information.
MCM builds a motif vocabulary in an unsupervised way and deploys a novel motif convolution operation to extract the local structural context.
arXiv Detail & Related papers (2022-08-09T03:37:37Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - TCL: Transformer-based Dynamic Graph Modelling via Contrastive Learning [87.38675639186405]
We propose a novel graph neural network approach, called TCL, which deals with the dynamically-evolving graph in a continuous-time fashion.
To the best of our knowledge, this is the first attempt to apply contrastive learning to representation learning on dynamic graphs.
arXiv Detail & Related papers (2021-05-17T15:33:25Z) - Structural Landmarking and Interaction Modelling: on Resolution Dilemmas
in Graph Classification [50.83222170524406]
We study the intrinsic difficulty in graph classification under the unified concept of resolution dilemmas''
We propose SLIM'', an inductive neural network model for Structural Landmarking and Interaction Modelling.
arXiv Detail & Related papers (2020-06-29T01:01:42Z) - Graph-based Visual-Semantic Entanglement Network for Zero-shot Image
Recognition [17.622748458955595]
We propose the Graph-based Visual-Semantic Entanglement Network to conduct graph modeling of visual features.
Our method outperforms state-of-the-art approaches on multiple representative ZSL datasets.
arXiv Detail & Related papers (2020-06-08T14:54:08Z) - Tensor Graph Convolutional Networks for Multi-relational and Robust
Learning [74.05478502080658]
This paper introduces a tensor-graph convolutional network (TGCN) for scalable semi-supervised learning (SSL) from data associated with a collection of graphs, that are represented by a tensor.
The proposed architecture achieves markedly improved performance relative to standard GCNs, copes with state-of-the-art adversarial attacks, and leads to remarkable SSL performance over protein-to-protein interaction networks.
arXiv Detail & Related papers (2020-03-15T02:33:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.