MWP-BERT: A Strong Baseline for Math Word Problems
- URL: http://arxiv.org/abs/2107.13435v1
- Date: Wed, 28 Jul 2021 15:28:41 GMT
- Title: MWP-BERT: A Strong Baseline for Math Word Problems
- Authors: Zhenwen Liang, Jipeng Zhang, Jie Shao, Xiangliang Zhang
- Abstract summary: Math word problem (MWP) solving is the task of transforming a sequence of natural language problem descriptions to executable math equations.
Although recent sequence modeling MWP solvers have gained credits on the math-text contextual understanding, pre-trained language models (PLM) have not been explored for solving MWP.
We introduce MWP-BERT to obtain pre-trained token representations that capture the alignment between text description and mathematical logic.
- Score: 47.51572465676904
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Math word problem (MWP) solving is the task of transforming a sequence of
natural language problem descriptions to executable math equations. An MWP
solver not only needs to understand complex scenarios described in the problem
texts, but also identify the key mathematical variables and associate text
descriptions with math equation logic. Although recent sequence modeling MWP
solvers have gained credits on the math-text contextual understanding,
pre-trained language models (PLM) have not been explored for solving MWP,
considering that PLM trained over free-form texts is limited in representing
text references to mathematical logic. In this work, we introduce MWP-BERT to
obtain pre-trained token representations that capture the alignment between
text description and mathematical logic. Additionally, we introduce a
keyword-based prompt matching method to address the MWPs requiring common-sense
knowledge. On a benchmark Math23K dataset and a new Ape210k dataset, we show
that MWP-BERT outperforms the strongest baseline model by 5-10% improvement on
accuracy.
Related papers
- Solving Math Word Problems with Reexamination [27.80592576792461]
We propose a pseudo-dual (PseDual) learning scheme to model such process, which is model-agnostic.
The pseudo-dual task is specifically defined as filling the numbers in the expression back into the original word problem with numbers masked.
Our pseudo-dual learning scheme has been tested and proven effective when being equipped in several representative MWP solvers through empirical studies.
arXiv Detail & Related papers (2023-10-14T14:23:44Z) - MWPRanker: An Expression Similarity Based Math Word Problem Retriever [12.638925774492403]
Math Word Problems (MWPs) in online assessments help test the ability of the learner to make critical inferences.
We propose a tool in this work for MWP retrieval.
arXiv Detail & Related papers (2023-07-03T15:44:18Z) - Learning by Analogy: Diverse Questions Generation in Math Word Problem [21.211970350827183]
Solving math word problem (MWP) with AI techniques has recently made great progress with the success of deep neural networks (DNN)
We argue that the ability of learning by analogy is essential for an MWP solver to better understand same problems which may typically be formulated in diverse ways.
In this paper, we make a first attempt to solve MWPs by generating diverse yet consistent questions/equations.
arXiv Detail & Related papers (2023-06-15T11:47:07Z) - Analogical Math Word Problems Solving with Enhanced Problem-Solution
Association [37.70402758178867]
We propose to build a novel MWP solver by leveraging analogical MWPs.
The key idea, named analogy identification, is to associate the analogical MWP pairs in a latent space.
A solution discriminator is integrated into the MWP solver to enhance the association between the representations of MWPs and their true solutions.
arXiv Detail & Related papers (2022-12-01T19:50:30Z) - Dynamic Prompt Learning via Policy Gradient for Semi-structured
Mathematical Reasoning [150.17907456113537]
We present Tabular Math Word Problems (TabMWP), a new dataset containing 38,431 grade-level problems that require mathematical reasoning.
We evaluate different pre-trained models on TabMWP, including the GPT-3 model in a few-shot setting.
We propose a novel approach, PromptPG, which utilizes policy gradient to learn to select in-context examples from a small amount of training data.
arXiv Detail & Related papers (2022-09-29T08:01:04Z) - LogicSolver: Towards Interpretable Math Word Problem Solving with
Logical Prompt-enhanced Learning [135.8654475934613]
We first construct a high-quality MWP dataset named InterMWP which consists of 11,495 MWPs.
We propose a novel approach with logical prompt and interpretation, called Logicr.
With these improved semantic representations, our Logicr generates corresponding solution expressions and interpretable knowledge in accord with the generated solution expressions.
arXiv Detail & Related papers (2022-05-17T11:01:52Z) - Unbiased Math Word Problems Benchmark for Mitigating Solving Bias [72.8677805114825]
Current solvers exist solving bias which consists of data bias and learning bias due to biased dataset and improper training strategy.
Our experiments verify MWP solvers are easy to be biased by the biased training datasets which do not cover diverse questions for each problem narrative of all MWPs.
An MWP can be naturally solved by multiple equivalent equations while current datasets take only one of the equivalent equations as ground truth.
arXiv Detail & Related papers (2022-05-17T06:07:04Z) - DISK: Domain-constrained Instance Sketch for Math Word Problem
Generation [16.045655800225436]
A math word problem (MWP) is a coherent narrative which reflects the underlying logic of math equations.
Previous methods mainly generate MWP text based on inflexible pre-defined templates.
We propose a neural model for generating MWP text from math equations.
arXiv Detail & Related papers (2022-04-10T13:54:23Z) - Generate & Rank: A Multi-task Framework for Math Word Problems [48.99880318686938]
Math word problem (MWP) is a challenging and critical task in natural language processing.
We propose Generate & Rank, a framework based on a generative pre-trained language model.
By joint training with generation and ranking, the model learns from its own mistakes and is able to distinguish between correct and incorrect expressions.
arXiv Detail & Related papers (2021-09-07T12:21:49Z) - Semantically-Aligned Universal Tree-Structured Solver for Math Word
Problems [129.90766822085132]
A practical automatic textual math word problems (MWPs) solver should be able to solve various textual MWPs.
We propose a simple but efficient method called Universal Expression Tree (UET) to make the first attempt to represent the equations of various MWPs uniformly.
Then a semantically-aligned universal tree-structured solver (SAU-r) based on an encoder-decoder framework is proposed to resolve multiple types of MWPs in a unified model.
arXiv Detail & Related papers (2020-10-14T06:27:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.