Solving Arithmetic Word Problems by Scoring Equations with Recursive
Neural Networks
- URL: http://arxiv.org/abs/2009.05639v2
- Date: Tue, 9 Mar 2021 12:49:35 GMT
- Title: Solving Arithmetic Word Problems by Scoring Equations with Recursive
Neural Networks
- Authors: Klim Zaporojets, Giannis Bekoulis, Johannes Deleu, Thomas Demeester,
Chris Develder
- Abstract summary: Recent works use automatic extraction and ranking of candidate solution equations providing the answer to arithmetic word problems.
In this work, we explore novel approaches to score such candidate solution equations using Tree-RNN configurations.
Our proposed method consists of transforming the mathematical expression of the equation into an expression tree.
- Score: 25.08023032443234
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Solving arithmetic word problems is a cornerstone task in assessing language
understanding and reasoning capabilities in NLP systems. Recent works use
automatic extraction and ranking of candidate solution equations providing the
answer to arithmetic word problems. In this work, we explore novel approaches
to score such candidate solution equations using tree-structured recursive
neural network (Tree-RNN) configurations. The advantage of this Tree-RNN
approach over using more established sequential representations, is that it can
naturally capture the structure of the equations. Our proposed method consists
of transforming the mathematical expression of the equation into an expression
tree. Further, we encode this tree into a Tree-RNN by using different Tree-LSTM
architectures. Experimental results show that our proposed method (i) improves
overall performance with more than 3% accuracy points compared to previous
state-of-the-art, and with over 15% points on a subset of problems that require
more complex reasoning, and (ii) outperforms sequential LSTMs by 4% accuracy
points on such more complex problems.
Related papers
- LiteSearch: Efficacious Tree Search for LLM [70.29796112457662]
This study introduces a novel guided tree search algorithm with dynamic node selection and node-level exploration budget.
Experiments conducted on the GSM8K and TabMWP datasets demonstrate that our approach enjoys significantly lower computational costs compared to baseline methods.
arXiv Detail & Related papers (2024-06-29T05:14:04Z) - TreeDQN: Learning to minimize Branch-and-Bound tree [78.52895577861327]
Branch-and-Bound is a convenient approach to solving optimization tasks in the form of Mixed Linear Programs.
The efficiency of the solver depends on the branchning used to select a variable for splitting.
We propose a reinforcement learning method that can efficiently learn the branching.
arXiv Detail & Related papers (2023-06-09T14:01:26Z) - Learning Branching Heuristics from Graph Neural Networks [1.4660170768702356]
We first propose a new graph neural network (GNN) model designed using a probabilistic method.
Our approach introduces a new way of applying GNNs towards enhancing the classical backtracking algorithm used in AI.
arXiv Detail & Related papers (2022-11-26T00:01:01Z) - A Recursively Recurrent Neural Network (R2N2) Architecture for Learning
Iterative Algorithms [64.3064050603721]
We generalize Runge-Kutta neural network to a recurrent neural network (R2N2) superstructure for the design of customized iterative algorithms.
We demonstrate that regular training of the weight parameters inside the proposed superstructure on input/output data of various computational problem classes yields similar iterations to Krylov solvers for linear equation systems, Newton-Krylov solvers for nonlinear equation systems, and Runge-Kutta solvers for ordinary differential equations.
arXiv Detail & Related papers (2022-11-22T16:30:33Z) - Learning to Prune Instances of Steiner Tree Problem in Graphs [0.47138177023764655]
We consider the Steiner tree problem on graphs where we are given a set of nodes.
The goal is to find a tree sub-graph that contains all nodes in the given set, potentially including additional nodes.
arXiv Detail & Related papers (2022-08-25T10:31:00Z) - What's Wrong with Deep Learning in Tree Search for Combinatorial
Optimization [8.879790406465556]
We present an open-source benchmark suite for the NP-hard Maximum Independent Set problem, in both its weighted and unweighted variants.
We also conduct an in-depth analysis of the popular guided tree search algorithm by Li et al. [NeurIPS 2018], testing various configurations on small and large synthetic and real-world graphs.
We show that the graph convolution network used in the tree search does not learn a meaningful representation of the solution structure, and can in fact be replaced by random values.
arXiv Detail & Related papers (2022-01-25T17:37:34Z) - Deterministic Iteratively Built KD-Tree with KNN Search for Exact
Applications [2.7325238096808318]
K-Nearest Neighbors (KNN) search is a fundamental algorithm in artificial intelligence software with applications in robotics, and autonomous vehicles.
Similar to binary trees, kd-trees become unbalanced as new data is added in online applications which can lead to rapid degradation in search performance unless the tree is rebuilt.
We will present a "forest of interval kd-trees" which reduces the number of tree rebuilds, without compromising the exactness of query results.
arXiv Detail & Related papers (2021-06-07T17:09:22Z) - Recognizing and Verifying Mathematical Equations using Multiplicative
Differential Neural Units [86.9207811656179]
We show that memory-augmented neural networks (NNs) can achieve higher-order, memory-augmented extrapolation, stable performance, and faster convergence.
Our models achieve a 1.53% average improvement over current state-of-the-art methods in equation verification and achieve a 2.22% Top-1 average accuracy and 2.96% Top-5 average accuracy for equation completion.
arXiv Detail & Related papers (2021-04-07T03:50:11Z) - Recursive Top-Down Production for Sentence Generation with Latent Trees [77.56794870399288]
We model the production property of context-free grammars for natural and synthetic languages.
We present a dynamic programming algorithm that marginalises over latent binary tree structures with $N$ leaves.
We also present experimental results on German-English translation on the Multi30k dataset.
arXiv Detail & Related papers (2020-10-09T17:47:16Z) - Strong Generalization and Efficiency in Neural Programs [69.18742158883869]
We study the problem of learning efficient algorithms that strongly generalize in the framework of neural program induction.
By carefully designing the input / output interfaces of the neural model and through imitation, we are able to learn models that produce correct results for arbitrary input sizes.
arXiv Detail & Related papers (2020-07-07T17:03:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.