Graph Agent: Explicit Reasoning Agent for Graphs
- URL: http://arxiv.org/abs/2310.16421v1
- Date: Wed, 25 Oct 2023 07:20:16 GMT
- Title: Graph Agent: Explicit Reasoning Agent for Graphs
- Authors: Qinyong Wang, Zhenxiang Gao, Rong Xu
- Abstract summary: We introduce the Graph Agent (GA), an intelligent agent methodology of leveraging large language models (LLMs), inductive-deductive reasoning modules, and long-term memory for graph reasoning tasks.
GA integrates aspects of symbolic reasoning and existing graph embedding methods to provide an innovative approach for complex graph reasoning tasks.
Results showed that GA reached state-of-the-art performance, demonstrating accuracy of 90.65%, 95.48%, and 89.32% on Cora, PubMed, and PrimeKG datasets, respectively.
- Score: 3.422149942031643
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Graph embedding methods such as Graph Neural Networks (GNNs) and Graph
Transformers have contributed to the development of graph reasoning algorithms
for various tasks on knowledge graphs. However, the lack of interpretability
and explainability of graph embedding methods has limited their applicability
in scenarios requiring explicit reasoning. In this paper, we introduce the
Graph Agent (GA), an intelligent agent methodology of leveraging large language
models (LLMs), inductive-deductive reasoning modules, and long-term memory for
knowledge graph reasoning tasks. GA integrates aspects of symbolic reasoning
and existing graph embedding methods to provide an innovative approach for
complex graph reasoning tasks. By converting graph structures into textual
data, GA enables LLMs to process, reason, and provide predictions alongside
human-interpretable explanations. The effectiveness of the GA was evaluated on
node classification and link prediction tasks. Results showed that GA reached
state-of-the-art performance, demonstrating accuracy of 90.65%, 95.48%, and
89.32% on Cora, PubMed, and PrimeKG datasets, respectively. Compared to
existing GNN and transformer models, GA offered advantages of explicit
reasoning ability, free-of-training, easy adaption to various graph reasoning
tasks
Related papers
- Graph-R1: Incentivizing the Zero-Shot Graph Learning Capability in LLMs via Explicit Reasoning [7.1931434571877375]
Graph Neural Networks (GNNs) are limited by fixed label spaces, while Large Language Models (LLMs) lack structural inductive biases.<n>Recent advances in Large Reasoning Models (LRMs) provide a zero-shot alternative via explicit, long chain-of-thought reasoning.<n>We propose a GNN-free approach that reformulates graph tasks--node classification, link prediction, and graph classification--as textual reasoning problems solved by LRMs.
arXiv Detail & Related papers (2025-08-24T14:49:02Z) - AugWard: Augmentation-Aware Representation Learning for Accurate Graph Classification [16.7104207718009]
AugWard is a graph representation learning framework that considers the diversity introduced by graph augmentation.
AugWard applies augmentation-aware training to predict the graph distance between the augmented graph and its original one.
Results show that AugWard gives the state-of-the-art performance in supervised, semi-supervised graph classification, and transfer learning.
arXiv Detail & Related papers (2025-03-27T02:58:28Z) - Revisiting Graph Neural Networks on Graph-level Tasks: Comprehensive Experiments, Analysis, and Improvements [54.006506479865344]
We propose a unified evaluation framework for graph-level Graph Neural Networks (GNNs)
This framework provides a standardized setting to evaluate GNNs across diverse datasets.
We also propose a novel GNN model with enhanced expressivity and generalization capabilities.
arXiv Detail & Related papers (2025-01-01T08:48:53Z) - The GECo algorithm for Graph Neural Networks Explanation [0.0]
We introduce a new methodology involving graph communities to address the interpretability of graph classification problems.
The proposed method, called GECo, exploits the idea that if a community is a subset of graph nodes densely connected, this property should play a role in graph classification.
The obtained results outperform the other methods for artificial graph datasets and most real-world datasets.
arXiv Detail & Related papers (2024-11-18T09:08:30Z) - GraphXAIN: Narratives to Explain Graph Neural Networks [0.0]
Graph Neural Networks (GNNs) are a powerful technique for machine learning on graph-structured data.
Existing GNN explanation methods yield technical outputs that are difficult for non-data scientists to understand.
We propose GraphXAIN, a method that generates natural language narratives explaining GNN predictions.
arXiv Detail & Related papers (2024-11-04T19:21:06Z) - Generating In-Distribution Proxy Graphs for Explaining Graph Neural Networks [17.71313964436965]
A popular paradigm for the explainability of GNNs is to identify explainable subgraphs by comparing their labels with the ones of original graphs.
This task is challenging due to the substantial distributional shift from the original graphs in the training set to the set of explainable subgraphs.
We propose a novel method that generates proxy graphs for explainable subgraphs that are in the distribution of training data.
arXiv Detail & Related papers (2024-02-03T05:19:02Z) - Graph Domain Adaptation: Challenges, Progress and Prospects [61.9048172631524]
We propose graph domain adaptation as an effective knowledge-transfer paradigm across graphs.
GDA introduces a bunch of task-related graphs as source graphs and adapts the knowledge learnt from source graphs to the target graphs.
We outline the research status and challenges, propose a taxonomy, introduce the details of representative works, and discuss the prospects.
arXiv Detail & Related papers (2024-02-01T02:44:32Z) - Interpretable Sparsification of Brain Graphs: Better Practices and
Effective Designs for Graph Neural Networks [15.101250958437038]
dense brain graphs pose computational challenges including high runtime and memory usage and limited interpretability.
We propose a new model, Interpretable Graph Sparsification, which enhances graph classification performance by up to 5.1% with 55.0% fewer edges.
arXiv Detail & Related papers (2023-06-26T01:37:10Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - MentorGNN: Deriving Curriculum for Pre-Training GNNs [61.97574489259085]
We propose an end-to-end model named MentorGNN that aims to supervise the pre-training process of GNNs across graphs.
We shed new light on the problem of domain adaption on relational data (i.e., graphs) by deriving a natural and interpretable upper bound on the generalization error of the pre-trained GNNs.
arXiv Detail & Related papers (2022-08-21T15:12:08Z) - Let Invariant Rationale Discovery Inspire Graph Contrastive Learning [98.10268114789775]
We argue that a high-performing augmentation should preserve the salient semantics of anchor graphs regarding instance-discrimination.
We propose a new framework, Rationale-aware Graph Contrastive Learning (RGCL)
RGCL uses a rationale generator to reveal salient features about graph instance-discrimination as the rationale, and then creates rationale-aware views for contrastive learning.
arXiv Detail & Related papers (2022-06-16T01:28:40Z) - ExplaGraphs: An Explanation Graph Generation Task for Structured
Commonsense Reasoning [65.15423587105472]
We present a new generative and structured commonsense-reasoning task (and an associated dataset) of explanation graph generation for stance prediction.
Specifically, given a belief and an argument, a model has to predict whether the argument supports or counters the belief and also generate a commonsense-augmented graph that serves as non-trivial, complete, and unambiguous explanation for the predicted stance.
A significant 83% of our graphs contain external commonsense nodes with diverse structures and reasoning depths.
arXiv Detail & Related papers (2021-04-15T17:51:36Z) - XGNN: Towards Model-Level Explanations of Graph Neural Networks [113.51160387804484]
Graphs neural networks (GNNs) learn node features by aggregating and combining neighbor information.
GNNs are mostly treated as black-boxes and lack human intelligible explanations.
We propose a novel approach, known as XGNN, to interpret GNNs at the model-level.
arXiv Detail & Related papers (2020-06-03T23:52:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.