Knowledge Enhanced Neural Networks for relational domains
- URL: http://arxiv.org/abs/2205.15762v1
- Date: Tue, 31 May 2022 13:00:34 GMT
- Title: Knowledge Enhanced Neural Networks for relational domains
- Authors: Alessandro Daniele, Luciano Serafini
- Abstract summary: We focus on a specific method, KENN, a Neural-Symbolic architecture that injects prior logical knowledge into a neural network.
In this paper, we propose an extension of KENN for relational data.
- Score: 83.9217787335878
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the recent past, there has been a growing interest in Neural-Symbolic
Integration frameworks, i.e., hybrid systems that integrate connectionist and
symbolic approaches to obtain the best of both worlds. In this work we focus on
a specific method, KENN (Knowledge Enhanced Neural Networks), a Neural-Symbolic
architecture that injects prior logical knowledge into a neural network by
adding on its top a residual layer that modifies the initial predictions
accordingly to the knowledge. Among the advantages of this strategy, there is
the inclusion of clause weights, learnable parameters that represent the
strength of the clauses, meaning that the model can learn the impact of each
rule on the final predictions. As a special case, if the training data
contradicts a constraint, KENN learns to ignore it, making the system robust to
the presence of wrong knowledge. In this paper, we propose an extension of KENN
for relational data. One of the main advantages of KENN resides in its
scalability, thanks to a flexible treatment of dependencies between the rules
obtained by stacking multiple logical layers. We show experimentally the
efficacy of this strategy. The results show that KENN is capable of increasing
the performances of the underlying neural network, obtaining better or
comparable accuracies in respect to other two related methods that combine
learning with logic, requiring significantly less time for learning.
Related papers
- GINN-KAN: Interpretability pipelining with applications in Physics Informed Neural Networks [5.2969467015867915]
We introduce the concept of interpretability pipelineing, to incorporate multiple interpretability techniques to outperform each individual technique.
We evaluate two recent models selected for their potential to incorporate interpretability into standard neural network architectures.
We introduce a novel interpretable neural network GINN-KAN that synthesizes the advantages of both models.
arXiv Detail & Related papers (2024-08-27T04:57:53Z) - BKDSNN: Enhancing the Performance of Learning-based Spiking Neural Networks Training with Blurred Knowledge Distillation [20.34272550256856]
Spiking neural networks (SNNs) mimic biological neural system to convey information via discrete spikes.
Our work achieves state-of-the-art performance for training SNNs on both static and neuromorphic datasets.
arXiv Detail & Related papers (2024-07-12T08:17:24Z) - Bayesian Neural Networks with Domain Knowledge Priors [52.80929437592308]
We propose a framework for integrating general forms of domain knowledge into a BNN prior.
We show that BNNs using our proposed domain knowledge priors outperform those with standard priors.
arXiv Detail & Related papers (2024-02-20T22:34:53Z) - A Hybrid Neural Coding Approach for Pattern Recognition with Spiking
Neural Networks [53.31941519245432]
Brain-inspired spiking neural networks (SNNs) have demonstrated promising capabilities in solving pattern recognition tasks.
These SNNs are grounded on homogeneous neurons that utilize a uniform neural coding for information representation.
In this study, we argue that SNN architectures should be holistically designed to incorporate heterogeneous coding schemes.
arXiv Detail & Related papers (2023-05-26T02:52:12Z) - Linear Leaky-Integrate-and-Fire Neuron Model Based Spiking Neural
Networks and Its Mapping Relationship to Deep Neural Networks [7.840247953745616]
Spiking neural networks (SNNs) are brain-inspired machine learning algorithms with merits such as biological plausibility and unsupervised learning capability.
This paper establishes a precise mathematical mapping between the biological parameters of the Linear Leaky-Integrate-and-Fire model (LIF)/SNNs and the parameters of ReLU-AN/Deep Neural Networks (DNNs)
arXiv Detail & Related papers (2022-05-31T17:02:26Z) - Reinforcement Learning with External Knowledge by using Logical Neural
Networks [67.46162586940905]
A recent neuro-symbolic framework called the Logical Neural Networks (LNNs) can simultaneously provide key-properties of both neural networks and symbolic logic.
We propose an integrated method that enables model-free reinforcement learning from external knowledge sources.
arXiv Detail & Related papers (2021-03-03T12:34:59Z) - Neural Networks Enhancement with Logical Knowledge [83.9217787335878]
We propose an extension of KENN for relational data.
The results show that KENN is capable of increasing the performances of the underlying neural network even in the presence relational data.
arXiv Detail & Related papers (2020-09-13T21:12:20Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.