Towards Tractable Mathematical Reasoning: Challenges, Strategies, and
Opportunities for Solving Math Word Problems
- URL: http://arxiv.org/abs/2111.05364v1
- Date: Fri, 29 Oct 2021 05:20:31 GMT
- Title: Towards Tractable Mathematical Reasoning: Challenges, Strategies, and
Opportunities for Solving Math Word Problems
- Authors: Keyur Faldu, Amit Sheth, Prashant Kikani, Manas Gaur, Aditi Avasthi
- Abstract summary: We inspect non-neural and neural methods to solve math word problems narrated in a natural language.
We highlight the ability of these methods to be generalizable, mathematically reasonable, interpretable, and explainable.
We discuss technological approaches, review the evolution of intuitive design choices to solve MWPs, and examine them for mathematical reasoning ability.
- Score: 4.309840398782996
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Mathematical reasoning would be one of the next frontiers for artificial
intelligence to make significant progress. The ongoing surge to solve math word
problems (MWPs) and hence achieve better mathematical reasoning ability would
continue to be a key line of research in the coming time. We inspect non-neural
and neural methods to solve math word problems narrated in a natural language.
We also highlight the ability of these methods to be generalizable,
mathematically reasonable, interpretable, and explainable. Neural approaches
dominate the current state of the art, and we survey them highlighting three
strategies to MWP solving: (1) direct answer generation, (2) expression tree
generation for inferring answers, and (3) template retrieval for answer
computation. Moreover, we discuss technological approaches, review the
evolution of intuitive design choices to solve MWPs, and examine them for
mathematical reasoning ability. We finally identify several gaps that warrant
the need for external knowledge and knowledge-infused learning, among several
other opportunities in solving MWPs.
Related papers
- Mathify: Evaluating Large Language Models on Mathematical Problem Solving Tasks [34.09857430966818]
We introduce an extensive mathematics dataset called "MathQuest" sourced from the 11th and 12th standard Mathematics NCERT textbooks.
We conduct fine-tuning experiments with three prominent large language models: LLaMA-2, WizardMath, and MAmmoTH.
Our experiments reveal that among the three models, MAmmoTH-13B emerges as the most proficient, achieving the highest level of competence in solving the presented mathematical problems.
arXiv Detail & Related papers (2024-04-19T08:45:42Z) - MathScale: Scaling Instruction Tuning for Mathematical Reasoning [70.89605383298331]
Large language models (LLMs) have demonstrated remarkable capabilities in problem-solving.
However, their proficiency in solving mathematical problems remains inadequate.
We propose MathScale, a simple and scalable method to create high-quality mathematical reasoning data.
arXiv Detail & Related papers (2024-03-05T11:42:59Z) - Do Language Models Exhibit the Same Cognitive Biases in Problem Solving as Human Learners? [140.9751389452011]
We study the biases of large language models (LLMs) in relation to those known in children when solving arithmetic word problems.
We generate a novel set of word problems for each of these tests, using a neuro-symbolic approach that enables fine-grained control over the problem features.
arXiv Detail & Related papers (2024-01-31T18:48:20Z) - Techniques to Improve Neural Math Word Problem Solvers [0.0]
Recent neural-based approaches mainly encode the problem text using a language model and decode a mathematical expression over quantities and operators iteratively.
We propose a new encoder-decoder architecture that fully leverages the question text and preserves step-wise commutative law.
Experiments on four established benchmarks demonstrate that our framework outperforms state-of-the-art neural MWP solvers.
arXiv Detail & Related papers (2023-02-06T22:41:51Z) - Towards a Holistic Understanding of Mathematical Questions with
Contrastive Pre-training [65.10741459705739]
We propose a novel contrastive pre-training approach for mathematical question representations, namely QuesCo.
We first design two-level question augmentations, including content-level and structure-level, which generate literally diverse question pairs with similar purposes.
Then, to fully exploit hierarchical information of knowledge concepts, we propose a knowledge hierarchy-aware rank strategy.
arXiv Detail & Related papers (2023-01-18T14:23:29Z) - A Survey of Deep Learning for Mathematical Reasoning [71.88150173381153]
We review the key tasks, datasets, and methods at the intersection of mathematical reasoning and deep learning over the past decade.
Recent advances in large-scale neural language models have opened up new benchmarks and opportunities to use deep learning for mathematical reasoning.
arXiv Detail & Related papers (2022-12-20T18:46:16Z) - JiuZhang: A Chinese Pre-trained Language Model for Mathematical Problem
Understanding [74.12405417718054]
This paper aims to advance the mathematical intelligence of machines by presenting the first Chinese mathematical pre-trained language model(PLM)
Unlike other standard NLP tasks, mathematical texts are difficult to understand, since they involve mathematical terminology, symbols and formulas in the problem statement.
We design a novel curriculum pre-training approach for improving the learning of mathematical PLMs, consisting of both basic and advanced courses.
arXiv Detail & Related papers (2022-06-13T17:03:52Z) - Why are NLP Models Fumbling at Elementary Math? A Survey of Deep
Learning based Word Problem Solvers [7.299537282917047]
We critically examine the various models that have been developed for solving word problems.
We take a step back and analyse why, in spite of this abundance in scholarly interest, the predominantly used experiment and dataset designs continue to be a stumbling block.
arXiv Detail & Related papers (2022-05-31T10:51:25Z) - Extracting the Unknown from Long Math Problems [8.19841678851784]
We propose computational methods for facilitating problem understanding through the task of recognizing the unknown in specifications of long Math problems.
Our experimental results show that learning models yield strong results on the task, a promising first step towards human interpretable, modular approaches to understanding long Math problems.
arXiv Detail & Related papers (2021-03-22T17:51:10Z) - Machine Number Sense: A Dataset of Visual Arithmetic Problems for
Abstract and Relational Reasoning [95.18337034090648]
We propose a dataset, Machine Number Sense (MNS), consisting of visual arithmetic problems automatically generated using a grammar model--And-Or Graph (AOG)
These visual arithmetic problems are in the form of geometric figures.
We benchmark the MNS dataset using four predominant neural network models as baselines in this visual reasoning task.
arXiv Detail & Related papers (2020-04-25T17:14:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.