Text Modular Networks: Learning to Decompose Tasks in the Language of
Existing Models
- URL: http://arxiv.org/abs/2009.00751v2
- Date: Mon, 12 Apr 2021 21:58:08 GMT
- Title: Text Modular Networks: Learning to Decompose Tasks in the Language of
Existing Models
- Authors: Tushar Khot and Daniel Khashabi and Kyle Richardson and Peter Clark
and Ashish Sabharwal
- Abstract summary: We propose a framework for building interpretable systems that learn to solve complex tasks by decomposing them into simpler ones solvable by existing models.
We use this framework to build ModularQA, a system that can answer multi-hop reasoning questions by decomposing them into sub-questions answerable by a neural factoid single-span QA model and a symbolic calculator.
- Score: 61.480085460269514
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a general framework called Text Modular Networks(TMNs) for
building interpretable systems that learn to solve complex tasks by decomposing
them into simpler ones solvable by existing models. To ensure solvability of
simpler tasks, TMNs learn the textual input-output behavior (i.e., language) of
existing models through their datasets. This differs from prior
decomposition-based approaches which, besides being designed specifically for
each complex task, produce decompositions independent of existing sub-models.
Specifically, we focus on Question Answering (QA) and show how to train a
next-question generator to sequentially produce sub-questions targeting
appropriate sub-models, without additional human annotation. These
sub-questions and answers provide a faithful natural language explanation of
the model's reasoning. We use this framework to build ModularQA, a system that
can answer multi-hop reasoning questions by decomposing them into sub-questions
answerable by a neural factoid single-span QA model and a symbolic calculator.
Our experiments show that ModularQA is more versatile than existing explainable
systems for DROP and HotpotQA datasets, is more robust than state-of-the-art
blackbox (uninterpretable) systems, and generates more understandable and
trustworthy explanations compared to prior work.
Related papers
- A Model Is Not Built By A Single Prompt: LLM-Based Domain Modeling With Question Decomposition [4.123601037699469]
In real-world domain modeling, engineers usually decompose complex tasks into easily solvable sub-tasks.
We propose an LLM-based domain modeling approach via question decomposition, similar to developer's modeling process.
Preliminary results show that our approach outperforms the single-prompt-based prompt.
arXiv Detail & Related papers (2024-10-13T14:28:04Z) - STOC-TOT: Stochastic Tree-of-Thought with Constrained Decoding for Complex Reasoning in Multi-Hop Question Answering [8.525847131940031]
Multi-hop question answering (MHQA) requires a model to retrieve and integrate information from multiple passages to answer a complex question.
Recent systems leverage the power of large language models and integrate evidence retrieval with reasoning prompts.
We propose STOC-TOT, a tree-of-thought reasoning prompting method with constrained decoding for MHQA.
arXiv Detail & Related papers (2024-07-04T07:17:53Z) - SEMQA: Semi-Extractive Multi-Source Question Answering [94.04430035121136]
We introduce a new QA task for answering multi-answer questions by summarizing multiple diverse sources in a semi-extractive fashion.
We create the first dataset of this kind, QuoteSum, with human-written semi-extractive answers to natural and generated questions.
arXiv Detail & Related papers (2023-11-08T18:46:32Z) - Modular Deep Learning [120.36599591042908]
Transfer learning has recently become the dominant paradigm of machine learning.
It remains unclear how to develop models that specialise towards multiple tasks without incurring negative interference.
Modular deep learning has emerged as a promising solution to these challenges.
arXiv Detail & Related papers (2023-02-22T18:11:25Z) - Successive Prompting for Decomposing Complex Questions [50.00659445976735]
Recent works leverage the capabilities of large language models (LMs) to perform complex question answering in a few-shot setting.
We introduce Successive Prompting'', where we iteratively break down a complex task into a simple task, solve it, and then repeat the process until we get the final solution.
Our best model (with successive prompting) achieves an improvement of 5% absolute F1 on a few-shot version of the DROP dataset.
arXiv Detail & Related papers (2022-12-08T06:03:38Z) - Hierarchical Conditional Relation Networks for Video Question Answering [62.1146543269993]
We introduce a general-purpose reusable neural unit called Conditional Relation Network (CRN)
CRN serves as a building block to construct more sophisticated structures for representation and reasoning over video.
Our evaluations on well-known datasets achieved new SoTA results, demonstrating the impact of building a general-purpose reasoning unit on complex domains such as VideoQA.
arXiv Detail & Related papers (2020-02-25T07:00:48Z) - ManyModalQA: Modality Disambiguation and QA over Diverse Inputs [73.93607719921945]
We present a new multimodal question answering challenge, ManyModalQA, in which an agent must answer a question by considering three distinct modalities.
We collect our data by scraping Wikipedia and then utilize crowdsourcing to collect question-answer pairs.
arXiv Detail & Related papers (2020-01-22T14:39:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.