Highlighting Named Entities in Input for Auto-Formulation of
Optimization Problems
- URL: http://arxiv.org/abs/2212.13201v3
- Date: Tue, 12 Dec 2023 17:18:48 GMT
- Title: Highlighting Named Entities in Input for Auto-Formulation of
Optimization Problems
- Authors: Neeraj Gangwar and Nickvash Kani
- Abstract summary: This paper presents an approach that converts linear programming word problems into mathematical formulations.
We leverage the named entities in the input and augment the input to highlight these entities.
Our approach achieves the highest accuracy among all submissions to the NL4Opt Competition, securing first place in the generation track.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Operations research deals with modeling and solving real-world problems as
mathematical optimization problems. While solving mathematical systems is
accomplished by analytical software, formulating a problem as a set of
mathematical operations has been typically done manually by domain experts.
Recent machine learning methods have shown promise in converting textual
problem descriptions to corresponding mathematical formulations. This paper
presents an approach that converts linear programming word problems into
mathematical formulations. We leverage the named entities in the input and
augment the input to highlight these entities. Our approach achieves the
highest accuracy among all submissions to the NL4Opt Competition, securing
first place in the generation track.
Related papers
- Autoformulation of Mathematical Optimization Models Using LLMs [50.030647274271516]
We develop an automated approach to creating optimization models from natural language descriptions for commercial solvers.
We identify the three core challenges of autoformulation: (1) defining the vast, problem-dependent hypothesis space, (2) efficiently searching this space under uncertainty, and (3) evaluating formulation correctness.
arXiv Detail & Related papers (2024-11-03T20:41:38Z) - Problem Categorization Can Help Large Language Models Solve Math Problems [0.0]
We show the effectiveness of using the classification of problems into different categories to facilitate problem-solving.
We also optimize the classification of problems into categories by creating an accurate dataset.
arXiv Detail & Related papers (2024-10-29T16:06:26Z) - Mathify: Evaluating Large Language Models on Mathematical Problem Solving Tasks [34.09857430966818]
We introduce an extensive mathematics dataset called "MathQuest" sourced from the 11th and 12th standard Mathematics NCERT textbooks.
We conduct fine-tuning experiments with three prominent large language models: LLaMA-2, WizardMath, and MAmmoTH.
Our experiments reveal that among the three models, MAmmoTH-13B emerges as the most proficient, achieving the highest level of competence in solving the presented mathematical problems.
arXiv Detail & Related papers (2024-04-19T08:45:42Z) - SEGO: Sequential Subgoal Optimization for Mathematical Problem-Solving [64.38649623473626]
Large Language Models (LLMs) have driven substantial progress in artificial intelligence.
We propose a novel framework called textbfSEquential subtextbfGoal textbfOptimization (SEGO) to enhance LLMs' ability to solve mathematical problems.
arXiv Detail & Related papers (2023-10-19T17:56:40Z) - Solving Math Word Problems by Combining Language Models With Symbolic
Solvers [28.010617102877923]
Large language models (LLMs) can be combined with external tools to perform complex reasoning and calculation.
We propose an approach that combines an LLM that can incrementally formalize word problems as a set of variables and equations with an external symbolic solver.
Our approach achieves comparable accuracy to the original PAL on the GSM8K benchmark of math word problems and outperforms PAL by an absolute 20% on ALGEBRA.
arXiv Detail & Related papers (2023-04-16T04:16:06Z) - JiuZhang: A Chinese Pre-trained Language Model for Mathematical Problem
Understanding [74.12405417718054]
This paper aims to advance the mathematical intelligence of machines by presenting the first Chinese mathematical pre-trained language model(PLM)
Unlike other standard NLP tasks, mathematical texts are difficult to understand, since they involve mathematical terminology, symbols and formulas in the problem statement.
We design a novel curriculum pre-training approach for improving the learning of mathematical PLMs, consisting of both basic and advanced courses.
arXiv Detail & Related papers (2022-06-13T17:03:52Z) - Tackling Math Word Problems with Fine-to-Coarse Abstracting and
Reasoning [22.127301797950572]
We propose to model a math word problem in a fine-to-coarse manner to capture both the local fine-grained information and the global logical structure of it.
Our model is naturally sensitive to local variations and can better generalize to unseen problem types.
arXiv Detail & Related papers (2022-05-17T12:14:44Z) - Measuring Mathematical Problem Solving With the MATH Dataset [55.4376028963537]
We introduce MATH, a dataset of 12,500 challenging competition mathematics problems.
Each problem has a full step-by-step solution which can be used to teach models to generate answer derivations and explanations.
We also contribute a large auxiliary pretraining dataset which helps teach models the fundamentals of mathematics.
arXiv Detail & Related papers (2021-03-05T18:59:39Z) - SMART: A Situation Model for Algebra Story Problems via Attributed
Grammar [74.1315776256292]
We introduce the concept of a emphsituation model, which originates from psychology studies to represent the mental states of humans in problem-solving.
We show that the proposed model outperforms all previous neural solvers by a large margin while preserving much better interpretability.
arXiv Detail & Related papers (2020-12-27T21:03:40Z) - Strong Generalization and Efficiency in Neural Programs [69.18742158883869]
We study the problem of learning efficient algorithms that strongly generalize in the framework of neural program induction.
By carefully designing the input / output interfaces of the neural model and through imitation, we are able to learn models that produce correct results for arbitrary input sizes.
arXiv Detail & Related papers (2020-07-07T17:03:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.