GraphQ IR: Unifying Semantic Parsing of Graph Query Language with
Intermediate Representation
- URL: http://arxiv.org/abs/2205.12078v1
- Date: Tue, 24 May 2022 13:59:53 GMT
- Title: GraphQ IR: Unifying Semantic Parsing of Graph Query Language with
Intermediate Representation
- Authors: Lunyiu Nie, Shulin Cao, Jiaxin Shi, Qi Tian, Lei Hou, Juanzi Li,
Jidong Zhai
- Abstract summary: We propose a unified intermediate representation (IR) for graph query languages, namely GraphQ IR.
With the IR's natural-language-like representation that bridges the semantic gap and its formally defined syntax that maintains the graph structure, neural semantic parsing can more effectively convert user queries into GraphQ IR.
Our approach can consistently achieve state-of-the-art performance on KQA Pro, Overnight and MetaQA.
- Score: 91.27083732371453
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Subject to the semantic gap lying between natural and formal language, neural
semantic parsing is typically bottlenecked by the paucity and imbalance of
data. In this paper, we propose a unified intermediate representation (IR) for
graph query languages, namely GraphQ IR. With the IR's natural-language-like
representation that bridges the semantic gap and its formally defined syntax
that maintains the graph structure, neural semantic parser can more effectively
convert user queries into our GraphQ IR, which can be later automatically
compiled into different downstream graph query languages. Extensive experiments
show that our approach can consistently achieve state-of-the-art performance on
benchmarks KQA Pro, Overnight and MetaQA. Evaluations under compositional
generalization and few-shot learning settings also validate the promising
generalization ability of GraphQ IR with at most 11% accuracy improvement.
Related papers
- Contextualized Messages Boost Graph Representations [1.5178009359320295]
This paper investigates the ability of graph networks (GNNs) to process data that may be represented as graphs.
It shows that only a few GNNs are investigated across all levels of capability.
A mathematical discussion on the relationship between SIRGCN and widely used GNNs is laid out to put the contribution into context.
arXiv Detail & Related papers (2024-03-19T08:05:49Z) - Conversational Semantic Parsing using Dynamic Context Graphs [68.72121830563906]
We consider the task of conversational semantic parsing over general purpose knowledge graphs (KGs) with millions of entities, and thousands of relation-types.
We focus on models which are capable of interactively mapping user utterances into executable logical forms.
arXiv Detail & Related papers (2023-05-04T16:04:41Z) - Learning to Count Isomorphisms with Graph Neural Networks [16.455234748896157]
Subgraph isomorphism counting is an important problem on graphs.
In this paper, we propose a novel graph neural network (GNN) called Count-GNN for subgraph isomorphism counting.
arXiv Detail & Related papers (2023-02-07T05:32:11Z) - Interactive Visual Pattern Search on Graph Data via Graph Representation
Learning [20.795511688640296]
We propose a visual analytics system GraphQ to support human-in-the-loop, example-based, subgraph pattern search.
To support fast, interactive queries, we use graph neural networks (GNNs) to encode a graph as fixed-length latent vector representation.
We also propose a novel GNN for node-alignment called NeuroAlign to facilitate easy validation and interpretation of the query results.
arXiv Detail & Related papers (2022-02-18T22:30:28Z) - GN-Transformer: Fusing Sequence and Graph Representation for Improved
Code Summarization [0.0]
We propose a novel method, GN-Transformer, to learn end-to-end on a fused sequence and graph modality.
The proposed methods achieve state-of-the-art performance in two code summarization datasets and across three automatic code summarization metrics.
arXiv Detail & Related papers (2021-11-17T02:51:37Z) - GraphiT: Encoding Graph Structure in Transformers [37.33808493548781]
We show that viewing graphs as sets of node features and structural and positional information is able to outperform representations learned with classical graph neural networks (GNNs)
Our model, GraphiT, encodes such information by (i) leveraging relative positional encoding strategies in self-attention scores based on positive definite kernels on graphs, and (ii) enumerating and encoding local sub-structures such as paths of short length.
arXiv Detail & Related papers (2021-06-10T11:36:22Z) - Learning the Implicit Semantic Representation on Graph-Structured Data [57.670106959061634]
Existing representation learning methods in graph convolutional networks are mainly designed by describing the neighborhood of each node as a perceptual whole.
We propose a Semantic Graph Convolutional Networks (SGCN) that explores the implicit semantics by learning latent semantic-paths in graphs.
arXiv Detail & Related papers (2021-01-16T16:18:43Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Iterative Context-Aware Graph Inference for Visual Dialog [126.016187323249]
We propose a novel Context-Aware Graph (CAG) neural network.
Each node in the graph corresponds to a joint semantic feature, including both object-based (visual) and history-related (textual) context representations.
arXiv Detail & Related papers (2020-04-05T13:09:37Z) - Evaluating Logical Generalization in Graph Neural Networks [59.70452462833374]
We study the task of logical generalization using graph neural networks (GNNs)
Our benchmark suite, GraphLog, requires that learning algorithms perform rule induction in different synthetic logics.
We find that the ability for models to generalize and adapt is strongly determined by the diversity of the logical rules they encounter during training.
arXiv Detail & Related papers (2020-03-14T05:45:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.