Learning by Analogy: Diverse Questions Generation in Math Word Problem
- URL: http://arxiv.org/abs/2306.09064v1
- Date: Thu, 15 Jun 2023 11:47:07 GMT
- Title: Learning by Analogy: Diverse Questions Generation in Math Word Problem
- Authors: Zihao Zhou, Maizhen Ning, Qiufeng Wang, Jie Yao, Wei Wang, Xiaowei
Huang, Kaizhu Huang
- Abstract summary: Solving math word problem (MWP) with AI techniques has recently made great progress with the success of deep neural networks (DNN)
We argue that the ability of learning by analogy is essential for an MWP solver to better understand same problems which may typically be formulated in diverse ways.
In this paper, we make a first attempt to solve MWPs by generating diverse yet consistent questions/equations.
- Score: 21.211970350827183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Solving math word problem (MWP) with AI techniques has recently made great
progress with the success of deep neural networks (DNN), but it is far from
being solved. We argue that the ability of learning by analogy is essential for
an MWP solver to better understand same problems which may typically be
formulated in diverse ways. However most existing works exploit the shortcut
learning to train MWP solvers simply based on samples with a single question.
In lack of diverse questions, these methods merely learn shallow heuristics. In
this paper, we make a first attempt to solve MWPs by generating diverse yet
consistent questions/equations. Given a typical MWP including the scenario
description, question, and equation (i.e., answer), we first generate multiple
consistent equations via a group of heuristic rules. We then feed them to a
question generator together with the scenario to obtain the corresponding
diverse questions, forming a new MWP with a variety of questions and equations.
Finally we engage a data filter to remove those unreasonable MWPs, keeping the
high-quality augmented ones. To evaluate the ability of learning by analogy for
an MWP solver, we generate a new MWP dataset (called DiverseMath23K) with
diverse questions by extending the current benchmark Math23K. Extensive
experimental results demonstrate that our proposed method can generate
high-quality diverse questions with corresponding equations, further leading to
performance improvement on Diverse-Math23K. The code and dataset is available
at: https://github.com/zhouzihao501/DiverseMWP
Related papers
- AI-Assisted Generation of Difficult Math Questions [78.7547836422727]
Current training positions mathematical reasoning as a core capability.
There is unmet demand for diverse and challenging math questions.
We present a design framework that combines the strengths of LLMs with a human-in-the-loop approach.
arXiv Detail & Related papers (2024-07-30T17:55:36Z) - Solving Math Word Problems with Reexamination [27.80592576792461]
We propose a pseudo-dual (PseDual) learning scheme to model such process, which is model-agnostic.
The pseudo-dual task is specifically defined as filling the numbers in the expression back into the original word problem with numbers masked.
Our pseudo-dual learning scheme has been tested and proven effective when being equipped in several representative MWP solvers through empirical studies.
arXiv Detail & Related papers (2023-10-14T14:23:44Z) - MWPRanker: An Expression Similarity Based Math Word Problem Retriever [12.638925774492403]
Math Word Problems (MWPs) in online assessments help test the ability of the learner to make critical inferences.
We propose a tool in this work for MWP retrieval.
arXiv Detail & Related papers (2023-07-03T15:44:18Z) - Generalizing Math Word Problem Solvers via Solution Diversification [56.2690023011738]
We design a new training framework for an MWP solver by introducing a solution buffer and a solution discriminator.
Our framework is flexibly applicable to a wide setting of fully, semi-weakly and weakly supervised training for all Seq2Seq MWP solvers.
arXiv Detail & Related papers (2022-12-01T19:34:58Z) - Dynamic Prompt Learning via Policy Gradient for Semi-structured
Mathematical Reasoning [150.17907456113537]
We present Tabular Math Word Problems (TabMWP), a new dataset containing 38,431 grade-level problems that require mathematical reasoning.
We evaluate different pre-trained models on TabMWP, including the GPT-3 model in a few-shot setting.
We propose a novel approach, PromptPG, which utilizes policy gradient to learn to select in-context examples from a small amount of training data.
arXiv Detail & Related papers (2022-09-29T08:01:04Z) - Unbiased Math Word Problems Benchmark for Mitigating Solving Bias [72.8677805114825]
Current solvers exist solving bias which consists of data bias and learning bias due to biased dataset and improper training strategy.
Our experiments verify MWP solvers are easy to be biased by the biased training datasets which do not cover diverse questions for each problem narrative of all MWPs.
An MWP can be naturally solved by multiple equivalent equations while current datasets take only one of the equivalent equations as ground truth.
arXiv Detail & Related papers (2022-05-17T06:07:04Z) - DISK: Domain-constrained Instance Sketch for Math Word Problem
Generation [16.045655800225436]
A math word problem (MWP) is a coherent narrative which reflects the underlying logic of math equations.
Previous methods mainly generate MWP text based on inflexible pre-defined templates.
We propose a neural model for generating MWP text from math equations.
arXiv Detail & Related papers (2022-04-10T13:54:23Z) - Math Word Problem Generation with Mathematical Consistency and Problem
Context Constraints [37.493809561634386]
We study the problem of generating arithmetic math word problems (MWPs) given a math equation.
Existing approaches are prone to generating MWPs that are mathematically invalid or have unsatisfactory language quality.
arXiv Detail & Related papers (2021-09-09T20:24:25Z) - Generate & Rank: A Multi-task Framework for Math Word Problems [48.99880318686938]
Math word problem (MWP) is a challenging and critical task in natural language processing.
We propose Generate & Rank, a framework based on a generative pre-trained language model.
By joint training with generation and ranking, the model learns from its own mistakes and is able to distinguish between correct and incorrect expressions.
arXiv Detail & Related papers (2021-09-07T12:21:49Z) - MWP-BERT: A Strong Baseline for Math Word Problems [47.51572465676904]
Math word problem (MWP) solving is the task of transforming a sequence of natural language problem descriptions to executable math equations.
Although recent sequence modeling MWP solvers have gained credits on the math-text contextual understanding, pre-trained language models (PLM) have not been explored for solving MWP.
We introduce MWP-BERT to obtain pre-trained token representations that capture the alignment between text description and mathematical logic.
arXiv Detail & Related papers (2021-07-28T15:28:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.