Heterogeneous Relational Reasoning in Knowledge Graphs with
Reinforcement Learning
- URL: http://arxiv.org/abs/2003.06050v1
- Date: Thu, 12 Mar 2020 22:39:58 GMT
- Title: Heterogeneous Relational Reasoning in Knowledge Graphs with
Reinforcement Learning
- Authors: Mandana Saebi, Steven Krieg, Chuxu Zhang, Meng Jiang, and Nitesh
Chawla
- Abstract summary: We introduce a type-enhanced reinforcement learning agent that uses the local neighborhood information for efficient path-based reasoning over knowledge graphs.
Our solution uses graph neural network (GNN) for encoding the neighborhood information and utilizes entity types to prune the action space.
- Score: 21.33973806169273
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Path-based relational reasoning over knowledge graphs has become increasingly
popular due to a variety of downstream applications such as question answering
in dialogue systems, fact prediction, and recommender systems. In recent years,
reinforcement learning (RL) has provided solutions that are more interpretable
and explainable than other deep learning models. However, these solutions still
face several challenges, including large action space for the RL agent and
accurate representation of entity neighborhood structure. We address these
problems by introducing a type-enhanced RL agent that uses the local
neighborhood information for efficient path-based reasoning over knowledge
graphs. Our solution uses graph neural network (GNN) for encoding the
neighborhood information and utilizes entity types to prune the action space.
Experiments on real-world dataset show that our method outperforms
state-of-the-art RL methods and discovers more novel paths during the training
procedure.
Related papers
- AriGraph: Learning Knowledge Graph World Models with Episodic Memory for LLM Agents [19.249596397679856]
AriGraph is a memory graph that integrates semantic and episodic memories while exploring the environment.
We demonstrate that our Ariadne LLM agent effectively handles complex tasks within interactive text game environments difficult even for human players.
arXiv Detail & Related papers (2024-07-05T09:06:47Z) - GASE: Graph Attention Sampling with Edges Fusion for Solving Vehicle Routing Problems [6.084414764415137]
We propose an adaptive Graph Attention Sampling with the Edges Fusion framework to solve vehicle routing problems.
Our proposed model outperforms the existing methods by 2.08%-6.23% and shows stronger generalization ability.
arXiv Detail & Related papers (2024-05-21T03:33:07Z) - Multi-modal Causal Structure Learning and Root Cause Analysis [67.67578590390907]
We propose Mulan, a unified multi-modal causal structure learning method for root cause localization.
We leverage a log-tailored language model to facilitate log representation learning, converting log sequences into time-series data.
We also introduce a novel key performance indicator-aware attention mechanism for assessing modality reliability and co-learning a final causal graph.
arXiv Detail & Related papers (2024-02-04T05:50:38Z) - Learning State-Augmented Policies for Information Routing in
Communication Networks [92.59624401684083]
We develop a novel State Augmentation (SA) strategy to maximize the aggregate information at source nodes using graph neural network (GNN) architectures.
We leverage an unsupervised learning procedure to convert the output of the GNN architecture to optimal information routing strategies.
In the experiments, we perform the evaluation on real-time network topologies to validate our algorithms.
arXiv Detail & Related papers (2023-09-30T04:34:25Z) - Iterative Zero-Shot LLM Prompting for Knowledge Graph Construction [104.29108668347727]
This paper proposes an innovative knowledge graph generation approach that leverages the potential of the latest generative large language models.
The approach is conveyed in a pipeline that comprises novel iterative zero-shot and external knowledge-agnostic strategies.
We claim that our proposal is a suitable solution for scalable and versatile knowledge graph construction and may be applied to different and novel contexts.
arXiv Detail & Related papers (2023-07-03T16:01:45Z) - Deep Reinforcement Learning Guided Improvement Heuristic for Job Shop
Scheduling [30.45126420996238]
This paper proposes a novel DRL-guided improvement for solving JSSP, where graph representation is employed to encode complete solutions.
We design a Graph Neural-Network-based representation scheme, consisting of two modules to effectively capture the information of dynamic topology and different types of nodes in graphs encountered during the improvement process.
We prove that our method scales linearly with problem size. Experiments on classic benchmarks show that the improvement policy learned by our method outperforms state-of-the-art DRL-based methods by a large margin.
arXiv Detail & Related papers (2022-11-20T10:20:13Z) - Agent-Controller Representations: Principled Offline RL with Rich
Exogenous Information [49.06422815335159]
Learning to control an agent from data collected offline is vital for real-world applications of reinforcement learning (RL)
This paper introduces offline RL benchmarks offering the ability to study this problem.
We find that contemporary representation learning techniques can fail on datasets where the noise is a complex and time dependent process.
arXiv Detail & Related papers (2022-10-31T22:12:48Z) - A Survey on Offline Reinforcement Learning: Taxonomy, Review, and Open
Problems [0.0]
Reinforcement learning (RL) has experienced a dramatic increase in popularity.
There is still a wide range of domains inaccessible to RL due to the high cost and danger of interacting with the environment.
offline RL is a paradigm that learns exclusively from static datasets of previously collected interactions.
arXiv Detail & Related papers (2022-03-02T20:05:11Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z) - Explainability in Deep Reinforcement Learning [68.8204255655161]
We review recent works in the direction to attain Explainable Reinforcement Learning (XRL)
In critical situations where it is essential to justify and explain the agent's behaviour, better explainability and interpretability of RL models could help gain scientific insight on the inner workings of what is still considered a black box.
arXiv Detail & Related papers (2020-08-15T10:11:42Z) - Graph-based State Representation for Deep Reinforcement Learning [1.5901689240516976]
We exploit the fact that the underlying Markov decision process (MDP) represents a graph, which enables us to incorporate the topological information for effective state representation learning.
Motivated by the recent success of node representations for several graph analytical tasks we specifically investigate the capability of node representation learning methods to effectively encode the topology of the underlying MDP in Deep RL.
We find that all embedding methods outperform the commonly used matrix representation of grid-world environments in all of the studied cases.
arXiv Detail & Related papers (2020-04-29T05:43:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.