Math Word Problem Solving by Generating Linguistic Variants of Problem
Statements
- URL: http://arxiv.org/abs/2306.13899v1
- Date: Sat, 24 Jun 2023 08:27:39 GMT
- Title: Math Word Problem Solving by Generating Linguistic Variants of Problem
Statements
- Authors: Syed Rifat Raiyan, Md. Nafis Faiyaz, Shah Md. Jawad Kabir, Mohsinul
Kabir, Hasan Mahmud, Md Kamrul Hasan
- Abstract summary: We propose a framework for MWP solvers based on the generation of linguistic variants of the problem text.
The approach involves solving each of the variant problems and electing the predicted expression with the majority of the votes.
We show that training on linguistic variants of problem statements and voting on candidate predictions improve the mathematical reasoning and robustness of the model.
- Score: 1.742186232261139
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The art of mathematical reasoning stands as a fundamental pillar of
intellectual progress and is a central catalyst in cultivating human ingenuity.
Researchers have recently published a plethora of works centered around the
task of solving Math Word Problems (MWP) $-$ a crucial stride towards general
AI. These existing models are susceptible to dependency on shallow heuristics
and spurious correlations to derive the solution expressions. In order to
ameliorate this issue, in this paper, we propose a framework for MWP solvers
based on the generation of linguistic variants of the problem text. The
approach involves solving each of the variant problems and electing the
predicted expression with the majority of the votes. We use DeBERTa
(Decoding-enhanced BERT with disentangled attention) as the encoder to leverage
its rich textual representations and enhanced mask decoder to construct the
solution expressions. Furthermore, we introduce a challenging dataset,
$\mathrm{P\small{ARA}\normalsize{MAWPS}}$, consisting of paraphrased,
adversarial, and inverse variants of selectively sampled MWPs from the
benchmark $\mathrm{M\small{AWPS}}$ dataset. We extensively experiment on this
dataset along with other benchmark datasets using some baseline MWP solver
models. We show that training on linguistic variants of problem statements and
voting on candidate predictions improve the mathematical reasoning and
robustness of the model. We make our code and data publicly available.
Related papers
- MWPRanker: An Expression Similarity Based Math Word Problem Retriever [12.638925774492403]
Math Word Problems (MWPs) in online assessments help test the ability of the learner to make critical inferences.
We propose a tool in this work for MWP retrieval.
arXiv Detail & Related papers (2023-07-03T15:44:18Z) - Techniques to Improve Neural Math Word Problem Solvers [0.0]
Recent neural-based approaches mainly encode the problem text using a language model and decode a mathematical expression over quantities and operators iteratively.
We propose a new encoder-decoder architecture that fully leverages the question text and preserves step-wise commutative law.
Experiments on four established benchmarks demonstrate that our framework outperforms state-of-the-art neural MWP solvers.
arXiv Detail & Related papers (2023-02-06T22:41:51Z) - Dynamic Prompt Learning via Policy Gradient for Semi-structured
Mathematical Reasoning [150.17907456113537]
We present Tabular Math Word Problems (TabMWP), a new dataset containing 38,431 grade-level problems that require mathematical reasoning.
We evaluate different pre-trained models on TabMWP, including the GPT-3 model in a few-shot setting.
We propose a novel approach, PromptPG, which utilizes policy gradient to learn to select in-context examples from a small amount of training data.
arXiv Detail & Related papers (2022-09-29T08:01:04Z) - Tackling Math Word Problems with Fine-to-Coarse Abstracting and
Reasoning [22.127301797950572]
We propose to model a math word problem in a fine-to-coarse manner to capture both the local fine-grained information and the global logical structure of it.
Our model is naturally sensitive to local variations and can better generalize to unseen problem types.
arXiv Detail & Related papers (2022-05-17T12:14:44Z) - LogicSolver: Towards Interpretable Math Word Problem Solving with
Logical Prompt-enhanced Learning [135.8654475934613]
We first construct a high-quality MWP dataset named InterMWP which consists of 11,495 MWPs.
We propose a novel approach with logical prompt and interpretation, called Logicr.
With these improved semantic representations, our Logicr generates corresponding solution expressions and interpretable knowledge in accord with the generated solution expressions.
arXiv Detail & Related papers (2022-05-17T11:01:52Z) - Unbiased Math Word Problems Benchmark for Mitigating Solving Bias [72.8677805114825]
Current solvers exist solving bias which consists of data bias and learning bias due to biased dataset and improper training strategy.
Our experiments verify MWP solvers are easy to be biased by the biased training datasets which do not cover diverse questions for each problem narrative of all MWPs.
An MWP can be naturally solved by multiple equivalent equations while current datasets take only one of the equivalent equations as ground truth.
arXiv Detail & Related papers (2022-05-17T06:07:04Z) - Seeking Patterns, Not just Memorizing Procedures: Contrastive Learning
for Solving Math Word Problems [14.144577791030853]
We investigate how a neural network understands patterns only from semantics.
We propose a contrastive learning approach, where the neural network perceives the divergence of patterns.
Our method greatly improves the performance in monolingual and multilingual settings.
arXiv Detail & Related papers (2021-10-16T04:03:47Z) - Generate & Rank: A Multi-task Framework for Math Word Problems [48.99880318686938]
Math word problem (MWP) is a challenging and critical task in natural language processing.
We propose Generate & Rank, a framework based on a generative pre-trained language model.
By joint training with generation and ranking, the model learns from its own mistakes and is able to distinguish between correct and incorrect expressions.
arXiv Detail & Related papers (2021-09-07T12:21:49Z) - MWP-BERT: A Strong Baseline for Math Word Problems [47.51572465676904]
Math word problem (MWP) solving is the task of transforming a sequence of natural language problem descriptions to executable math equations.
Although recent sequence modeling MWP solvers have gained credits on the math-text contextual understanding, pre-trained language models (PLM) have not been explored for solving MWP.
We introduce MWP-BERT to obtain pre-trained token representations that capture the alignment between text description and mathematical logic.
arXiv Detail & Related papers (2021-07-28T15:28:41Z) - Learning by Fixing: Solving Math Word Problems with Weak Supervision [70.62896781438694]
Previous neural solvers of math word problems (MWPs) are learned with full supervision and fail to generate diverse solutions.
We introduce a textitweakly-supervised paradigm for learning MWPs.
Our method only requires the annotations of the final answers and can generate various solutions for a single problem.
arXiv Detail & Related papers (2020-12-19T03:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.