Learning Division with Neural Arithmetic Logic Modules
- URL: http://arxiv.org/abs/2110.05177v2
- Date: Tue, 12 Oct 2021 08:18:00 GMT
- Title: Learning Division with Neural Arithmetic Logic Modules
- Authors: Bhumika Mistry, Katayoun Farrahi, Jonathon Hare
- Abstract summary: We show that robustly learning division in a systematic manner remains a challenge even at the simplest level of dividing two numbers.
We propose two novel approaches for division which we call the Neural Reciprocal Unit (NRU) and the Neural Multiplicative Reciprocal Unit (NMRU)
- Score: 2.019622939313173
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To achieve systematic generalisation, it first makes sense to master simple
tasks such as arithmetic. Of the four fundamental arithmetic operations
(+,-,$\times$,$\div$), division is considered the most difficult for both
humans and computers. In this paper we show that robustly learning division in
a systematic manner remains a challenge even at the simplest level of dividing
two numbers. We propose two novel approaches for division which we call the
Neural Reciprocal Unit (NRU) and the Neural Multiplicative Reciprocal Unit
(NMRU), and present improvements for an existing division module, the Real
Neural Power Unit (Real NPU). Experiments in learning division with input
redundancy on 225 different training sets, find that our proposed modifications
to the Real NPU obtains an average success of 85.3$\%$ improving over the
original by 15.1$\%$. In light of the suggestion above, our NMRU approach can
further improve the success to 91.6$\%$.
Related papers
- Reverse That Number! Decoding Order Matters in Arithmetic Learning [49.5504492920404]
Our work introduces a novel strategy that reevaluates the digit order by prioritizing output from the least significant digit.
Compared to the previous state-of-the-art (SOTA) method, our findings reveal an overall improvement of in accuracy while requiring only a third of the tokens typically used during training.
arXiv Detail & Related papers (2024-03-09T09:04:53Z) - RevOrder: A Novel Method for Enhanced Arithmetic in Language Models [0.9043578619916238]
RevOrder reverses the output digits in addition, subtraction, and n-digit by 1-digit (nD by 1D) multiplication tasks.
Our method significantly reduces the Count of Sequential Intermediate Digits (CSID) to $mathcalO(1)$.
arXiv Detail & Related papers (2024-02-06T09:10:35Z) - Decouple Graph Neural Networks: Train Multiple Simple GNNs Simultaneously Instead of One [60.5818387068983]
Graph neural networks (GNN) suffer from severe inefficiency.
We propose to decouple a multi-layer GNN as multiple simple modules for more efficient training.
We show that the proposed framework is highly efficient with reasonable performance.
arXiv Detail & Related papers (2023-04-20T07:21:32Z) - Memorization Capacity of Neural Networks with Conditional Computation [14.048989759890475]
We show that memorizing a collection of $n$ input-output relationships can be accomplished via a neural network with $O(sqrtn)$ neurons.
Using a conditional ReLU network, we show that the same task can be accomplished using only $O(log n)$ operations per input.
arXiv Detail & Related papers (2023-03-20T16:33:17Z) - Training Data is More Valuable than You Think: A Simple and Effective
Method by Retrieving from Training Data [82.92758444543689]
Retrieval-based methods have been shown to be effective in NLP tasks via introducing external knowledge.
Surprisingly, we found that REtrieving from the traINing datA (REINA) only can lead to significant gains on multiple NLG and NLU tasks.
Experimental results show that this simple method can achieve significantly better performance on a variety of NLU and NLG tasks.
arXiv Detail & Related papers (2022-03-16T17:37:27Z) - Self-Ensembling GAN for Cross-Domain Semantic Segmentation [107.27377745720243]
This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
arXiv Detail & Related papers (2021-12-15T09:50:25Z) - Recognizing and Verifying Mathematical Equations using Multiplicative
Differential Neural Units [86.9207811656179]
We show that memory-augmented neural networks (NNs) can achieve higher-order, memory-augmented extrapolation, stable performance, and faster convergence.
Our models achieve a 1.53% average improvement over current state-of-the-art methods in equation verification and achieve a 2.22% Top-1 average accuracy and 2.96% Top-5 average accuracy for equation completion.
arXiv Detail & Related papers (2021-04-07T03:50:11Z) - Neural Power Units [1.7188280334580197]
We introduce the Neural Power Unit (NPU) that operates on the full domain of real numbers and is capable of learning arbitrary power functions in a single layer.
We show that the NPUs outperform their competitors in terms of accuracy and sparsity on artificial arithmetic datasets.
arXiv Detail & Related papers (2020-06-02T14:58:07Z) - iNALU: Improved Neural Arithmetic Logic Unit [2.331160520377439]
The recently proposed Neural Arithmetic Logic Unit (NALU) is a novel neural architecture which is able to explicitly represent the mathematical relationships by the units of the network to learn operations such as summation, subtraction or multiplication.
We show that our model solves stability issues and outperforms the original NALU model in means of arithmetic precision and convergence.
arXiv Detail & Related papers (2020-03-17T10:37:22Z) - Neural Arithmetic Units [84.65228064780744]
Neural networks can approximate complex functions, but they struggle to perform exact arithmetic operations over real numbers.
We present two new neural network components: the Neural Addition Unit (NAU), which can learn exact addition and subtraction, and the Neural multiplication Unit (NMU), which can multiply subsets of a vector.
Our proposed units NAU and NMU, compared with previous neural units, converge more consistently, have fewer parameters, learn faster, can converge for larger hidden sizes, obtain sparse and meaningful weights, and can extrapolate to negative and small values.
arXiv Detail & Related papers (2020-01-14T19:35:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.