ULLER: A Unified Language for Learning and Reasoning
- URL: http://arxiv.org/abs/2405.00532v3
- Date: Wed, 3 Jul 2024 06:34:31 GMT
- Title: ULLER: A Unified Language for Learning and Reasoning
- Authors: Emile van Krieken, Samy Badreddine, Robin Manhaeve, Eleonora Giunchiglia,
- Abstract summary: We propose a unified language for neuro-symbolic artificial intelligence (NeSy)
We call it ULLER, a Unified Language for LEarning and Reasoning.
- Score: 7.689000976615671
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The field of neuro-symbolic artificial intelligence (NeSy), which combines learning and reasoning, has recently experienced significant growth. There now are a wide variety of NeSy frameworks, each with its own specific language for expressing background knowledge and how to relate it to neural networks. This heterogeneity hinders accessibility for newcomers and makes comparing different NeSy frameworks challenging. We propose a unified language for NeSy, which we call ULLER, a Unified Language for LEarning and Reasoning. ULLER encompasses a wide variety of settings, while ensuring that knowledge described in it can be used in existing NeSy systems. ULLER has a neuro-symbolic first-order syntax for which we provide example semantics including classical, fuzzy, and probabilistic logics. We believe ULLER is a first step towards making NeSy research more accessible and comparable, paving the way for libraries that streamline training and evaluation across a multitude of semantics, knowledge bases, and NeSy systems.
Related papers
- A Mathematical Framework, a Taxonomy of Modeling Paradigms, and a Suite of Learning Techniques for Neural-Symbolic Systems [22.42431063362667]
We introduce Neural-Symbolic Energy-Based Models (NeSy-EBMs), a unifying mathematical framework for discnative and generative modeling.
We utilize NeSy-EBMs to develop a taxonomy of modeling paradigms focusing on a system's neural-symbolic interface and reasoning capabilities.
We also present Neural Probabilistic Soft Logic (NeuPSL), an open-source NeSy-EBM library designed for scalability and expressivity.
arXiv Detail & Related papers (2024-07-12T21:26:21Z) - Weakly Supervised Reasoning by Neuro-Symbolic Approaches [28.98845133698169]
We will introduce our progress on neuro-symbolic approaches to NLP.
We will design a neural system with symbolic latent structures for an NLP task.
We will apply reinforcement learning or its relaxation to perform weakly supervised reasoning in the downstream task.
arXiv Detail & Related papers (2023-09-19T06:10:51Z) - Mindstorms in Natural Language-Based Societies of Mind [110.05229611910478]
Minsky's "society of mind" and Schmidhuber's "learning to think" inspire diverse societies of large multimodal neural networks (NNs)
Recent implementations of NN-based societies of minds consist of large language models (LLMs) and other NN-based experts communicating through a natural language interface.
In these natural language-based societies of mind (NLSOMs), new agents -- all communicating through the same universal symbolic language -- are easily added in a modular fashion.
arXiv Detail & Related papers (2023-05-26T16:21:25Z) - Multi-level Contrastive Learning for Cross-lingual Spoken Language
Understanding [90.87454350016121]
We develop novel code-switching schemes to generate hard negative examples for contrastive learning at all levels.
We develop a label-aware joint model to leverage label semantics for cross-lingual knowledge transfer.
arXiv Detail & Related papers (2022-05-07T13:44:28Z) - MRKL Systems: A modular, neuro-symbolic architecture that combines large
language models, external knowledge sources and discrete reasoning [50.40151403246205]
Huge language models (LMs) have ushered in a new era for AI, serving as a gateway to natural-language-based knowledge tasks.
We define a flexible architecture with multiple neural models, complemented by discrete knowledge and reasoning modules.
We describe this neuro-symbolic architecture, dubbed the Modular Reasoning, Knowledge and Language (MRKL) system.
arXiv Detail & Related papers (2022-05-01T11:01:28Z) - Is Neuro-Symbolic AI Meeting its Promise in Natural Language Processing?
A Structured Review [2.064612766965483]
Advocates for Neuro-Symbolic AI (NeSy) assert that combining deep learning with symbolic reasoning will lead to stronger AI.
We conduct a structured review of studies implementing NeSy for NLP, challenges and future directions.
We aim to answer the question of whether NeSy is indeed meeting its promises: reasoning, out-of-distribution generalization, interpretability, learning and reasoning from small data, and transferability to new domains.
arXiv Detail & Related papers (2022-02-24T17:13:33Z) - Emergence of Machine Language: Towards Symbolic Intelligence with Neural
Networks [73.94290462239061]
We propose to combine symbolism and connectionism principles by using neural networks to derive a discrete representation.
By designing an interactive environment and task, we demonstrated that machines could generate a spontaneous, flexible, and semantic language.
arXiv Detail & Related papers (2022-01-14T14:54:58Z) - Towards Zero-shot Language Modeling [90.80124496312274]
We construct a neural model that is inductively biased towards learning human languages.
We infer this distribution from a sample of typologically diverse training languages.
We harness additional language-specific side information as distant supervision for held-out languages.
arXiv Detail & Related papers (2021-08-06T23:49:18Z) - NeuralLog: Natural Language Inference with Joint Neural and Logical
Reasoning [6.795509403707242]
We propose an inference framework called NeuralLog, which utilizes both a monotonicity-based logical inference engine and a neural network language model for phrase alignment.
Our framework models the NLI task as a classic search problem and uses the beam search algorithm to search for optimal inference paths.
Experiments show that our joint logic and neural inference system improves accuracy on the NLI task and can achieve state-of-art accuracy on the SICK and MED datasets.
arXiv Detail & Related papers (2021-05-29T01:02:40Z) - Neural Networks Enhancement with Logical Knowledge [83.9217787335878]
We propose an extension of KENN for relational data.
The results show that KENN is capable of increasing the performances of the underlying neural network even in the presence relational data.
arXiv Detail & Related papers (2020-09-13T21:12:20Z) - Compositional Languages Emerge in a Neural Iterated Learning Model [27.495624644227888]
compositionality enables natural language to represent complex concepts via a structured combination of simpler ones.
We propose an effective neural iterated learning (NIL) algorithm that, when applied to interacting neural agents, facilitates the emergence of a more structured type of language.
arXiv Detail & Related papers (2020-02-04T15:19:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.