Neural-Symbolic Solver for Math Word Problems with Auxiliary Tasks
- URL: http://arxiv.org/abs/2107.01431v1
- Date: Sat, 3 Jul 2021 13:14:58 GMT
- Title: Neural-Symbolic Solver for Math Word Problems with Auxiliary Tasks
- Authors: Jinghui Qin, Xiaodan Liang, Yining Hong, Jianheng Tang, Liang Lin
- Abstract summary: Our NS-r consists of a problem reader to encode problems, a programmer to generate symbolic equations, and a symbolic executor to obtain answers.
Along with target expression supervision, our solver is also optimized via 4 new auxiliary objectives to enforce different symbolic reasoning.
- Score: 130.70449023574537
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Previous math word problem solvers following the encoder-decoder paradigm
fail to explicitly incorporate essential math symbolic constraints, leading to
unexplainable and unreasonable predictions. Herein, we propose Neural-Symbolic
Solver (NS-Solver) to explicitly and seamlessly incorporate different levels of
symbolic constraints by auxiliary tasks. Our NS-Solver consists of a problem
reader to encode problems, a programmer to generate symbolic equations, and a
symbolic executor to obtain answers. Along with target expression supervision,
our solver is also optimized via 4 new auxiliary objectives to enforce
different symbolic reasoning: a) self-supervised number prediction task
predicting both number quantity and number locations; b) commonsense constant
prediction task predicting what prior knowledge (e.g. how many legs a chicken
has) is required; c) program consistency checker computing the semantic loss
between predicted equation and target equation to ensure reasonable equation
mapping; d) duality exploiting task exploiting the quasi duality between
symbolic equation generation and problem's part-of-speech generation to enhance
the understanding ability of a solver. Besides, to provide a more realistic and
challenging benchmark for developing a universal and scalable solver, we also
construct a new large-scale MWP benchmark CM17K consisting of 4 kinds of MWPs
(arithmetic, one-unknown linear, one-unknown non-linear, equation set) with
more than 17K samples. Extensive experiments on Math23K and our CM17k
demonstrate the superiority of our NS-Solver compared to state-of-the-art
methods.
Related papers
- An Unsupervised Learning Framework Combined with Heuristics for the Maximum Minimal Cut Problem [5.092968949752308]
This work proposes an unsupervised learning framework combined with Maximums for MMCP.
A crucial observation is that each solution corresponds to at least one spanning tree.
We conduct extensive experiments to evaluate our framework and give a specific application.
arXiv Detail & Related papers (2024-08-16T02:07:34Z) - Techniques to Improve Neural Math Word Problem Solvers [0.0]
Recent neural-based approaches mainly encode the problem text using a language model and decode a mathematical expression over quantities and operators iteratively.
We propose a new encoder-decoder architecture that fully leverages the question text and preserves step-wise commutative law.
Experiments on four established benchmarks demonstrate that our framework outperforms state-of-the-art neural MWP solvers.
arXiv Detail & Related papers (2023-02-06T22:41:51Z) - On Robust Numerical Solver for ODE via Self-Attention Mechanism [82.95493796476767]
We explore training efficient and robust AI-enhanced numerical solvers with a small data size by mitigating intrinsic noise disturbances.
We first analyze the ability of the self-attention mechanism to regulate noise in supervised learning and then propose a simple-yet-effective numerical solver, Attr, which introduces an additive self-attention mechanism to the numerical solution of differential equations.
arXiv Detail & Related papers (2023-02-05T01:39:21Z) - Seeking Diverse Reasoning Logic: Controlled Equation Expression
Generation for Solving Math Word Problems [21.62131402402428]
We propose a controlled equation generation solver by leveraging a set of control codes to guide the model.
Our method universally improves the performance on single-unknown (Math23K) and multiple-unknown (DRAW1K, HMWP) benchmarks.
arXiv Detail & Related papers (2022-09-21T12:43:30Z) - Recognizing and Verifying Mathematical Equations using Multiplicative
Differential Neural Units [86.9207811656179]
We show that memory-augmented neural networks (NNs) can achieve higher-order, memory-augmented extrapolation, stable performance, and faster convergence.
Our models achieve a 1.53% average improvement over current state-of-the-art methods in equation verification and achieve a 2.22% Top-1 average accuracy and 2.96% Top-5 average accuracy for equation completion.
arXiv Detail & Related papers (2021-04-07T03:50:11Z) - SMART: A Situation Model for Algebra Story Problems via Attributed
Grammar [74.1315776256292]
We introduce the concept of a emphsituation model, which originates from psychology studies to represent the mental states of humans in problem-solving.
We show that the proposed model outperforms all previous neural solvers by a large margin while preserving much better interpretability.
arXiv Detail & Related papers (2020-12-27T21:03:40Z) - Closed Loop Neural-Symbolic Learning via Integrating Neural Perception,
Grammar Parsing, and Symbolic Reasoning [134.77207192945053]
Prior methods learn the neural-symbolic models using reinforcement learning approaches.
We introduce the textbfgrammar model as a textitsymbolic prior to bridge neural perception and symbolic reasoning.
We propose a novel textbfback-search algorithm which mimics the top-down human-like learning procedure to propagate the error.
arXiv Detail & Related papers (2020-06-11T17:42:49Z) - Machine Number Sense: A Dataset of Visual Arithmetic Problems for
Abstract and Relational Reasoning [95.18337034090648]
We propose a dataset, Machine Number Sense (MNS), consisting of visual arithmetic problems automatically generated using a grammar model--And-Or Graph (AOG)
These visual arithmetic problems are in the form of geometric figures.
We benchmark the MNS dataset using four predominant neural network models as baselines in this visual reasoning task.
arXiv Detail & Related papers (2020-04-25T17:14:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.