A Primer for Neural Arithmetic Logic Modules
- URL: http://arxiv.org/abs/2101.09530v2
- Date: Mon, 8 Aug 2022 15:04:24 GMT
- Title: A Primer for Neural Arithmetic Logic Modules
- Authors: Bhumika Mistry, Katayoun Farrahi and Jonathon Hare
- Abstract summary: This paper is the first in discussing the current state of progress of this field.
Focusing on the shortcomings of the NALU, we provide an in-depth analysis to reason about design choices of recent modules.
To alleviate the existing inconsistencies, we create a benchmark which compares all existing arithmetic NALMs.
- Score: 2.4278445972594525
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural Arithmetic Logic Modules have become a growing area of interest,
though remain a niche field. These modules are neural networks which aim to
achieve systematic generalisation in learning arithmetic and/or logic
operations such as $\{+, -, \times, \div, \leq, \textrm{AND}\}$ while also
being interpretable. This paper is the first in discussing the current state of
progress of this field, explaining key works, starting with the Neural
Arithmetic Logic Unit (NALU). Focusing on the shortcomings of the NALU, we
provide an in-depth analysis to reason about design choices of recent modules.
A cross-comparison between modules is made on experiment setups and findings,
where we highlight inconsistencies in a fundamental experiment causing the
inability to directly compare across papers. To alleviate the existing
inconsistencies, we create a benchmark which compares all existing arithmetic
NALMs. We finish by providing a novel discussion of existing applications for
NALU and research directions requiring further exploration.
Related papers
- Improving Complex Reasoning over Knowledge Graph with Logic-Aware Curriculum Tuning [89.89857766491475]
We propose a complex reasoning schema over KG upon large language models (LLMs)
We augment the arbitrary first-order logical queries via binary tree decomposition to stimulate the reasoning capability of LLMs.
Experiments across widely used datasets demonstrate that LACT has substantial improvements(brings an average +5.5% MRR score) over advanced methods.
arXiv Detail & Related papers (2024-05-02T18:12:08Z) - logLTN: Differentiable Fuzzy Logic in the Logarithm Space [11.440949097704943]
A trend in the literature involves integrating axioms and facts in loss functions by grounding logical symbols with neural networks and fuzzy semantics.
This paper presents a configuration of fuzzy operators for grounding formulas end-to-end in the logarithm space.
Our findings, both formal and empirical, show that the proposed configuration outperforms the state-of-the-art.
arXiv Detail & Related papers (2023-06-26T09:39:05Z) - Teaching Neural Module Networks to Do Arithmetic [54.06832128723388]
We up-grade NMNs by bridging the gap between its interpreter and the complex questions.
We introduce addition and subtraction modules that perform numerical reasoning over numbers.
On a subset of DROP, experimental results show that our proposed methods enhance NMNs' numerical reasoning skills by 17.7% improvement of F1 score.
arXiv Detail & Related papers (2022-10-06T06:38:04Z) - Linear Temporal Logic Modulo Theories over Finite Traces (Extended
Version) [72.38188258853155]
This paper studies Linear Temporal Logic over Finite Traces (LTLf)
proposition letters are replaced with first-order formulas interpreted over arbitrary theories.
The resulting logic, called Satisfiability Modulo Theories (LTLfMT), is semi-decidable.
arXiv Detail & Related papers (2022-04-28T17:57:33Z) - Neural Sequence-to-grid Module for Learning Symbolic Rules [14.946594806351971]
We propose a neural sequence-to-grid (seq2grid) module, an input preprocessor that automatically segments and aligns an input sequence into a grid.
As our module outputs a grid via a novel differentiable mapping, any neural network structure taking a grid input, such as ResNet or TextCNN, can be jointly trained with our module.
arXiv Detail & Related papers (2021-01-13T07:53:14Z) - Neural Function Modules with Sparse Arguments: A Dynamic Approach to
Integrating Information across Layers [84.57980167400513]
Neural Function Modules (NFM) aims to introduce the same structural capability into deep learning.
Most of the work in the context of feed-forward networks combining top-down and bottom-up feedback is limited to classification problems.
The key contribution of our work is to combine attention, sparsity, top-down and bottom-up feedback, in a flexible algorithm.
arXiv Detail & Related papers (2020-10-15T20:43:17Z) - Logical Natural Language Generation from Open-Domain Tables [107.04385677577862]
We propose a new task where a model is tasked with generating natural language statements that can be emphlogically entailed by the facts.
To facilitate the study of the proposed logical NLG problem, we use the existing TabFact dataset citechen 2019tabfact featured with a wide range of logical/symbolic inferences.
The new task poses challenges to the existing monotonic generation frameworks due to the mismatch between sequence order and logical order.
arXiv Detail & Related papers (2020-04-22T06:03:10Z) - iNALU: Improved Neural Arithmetic Logic Unit [2.331160520377439]
The recently proposed Neural Arithmetic Logic Unit (NALU) is a novel neural architecture which is able to explicitly represent the mathematical relationships by the units of the network to learn operations such as summation, subtraction or multiplication.
We show that our model solves stability issues and outperforms the original NALU model in means of arithmetic precision and convergence.
arXiv Detail & Related papers (2020-03-17T10:37:22Z) - Evaluating Logical Generalization in Graph Neural Networks [59.70452462833374]
We study the task of logical generalization using graph neural networks (GNNs)
Our benchmark suite, GraphLog, requires that learning algorithms perform rule induction in different synthetic logics.
We find that the ability for models to generalize and adapt is strongly determined by the diversity of the logical rules they encounter during training.
arXiv Detail & Related papers (2020-03-14T05:45:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.