Ada-Instruct: Adapting Instruction Generators for Complex Reasoning
- URL: http://arxiv.org/abs/2310.04484v2
- Date: Tue, 10 Oct 2023 07:17:32 GMT
- Title: Ada-Instruct: Adapting Instruction Generators for Complex Reasoning
- Authors: Wanyun Cui, Qianle Wang
- Abstract summary: We introduce Ada-Instruct, an adaptive instruction generator developed by fine-tuning open-source LLMs.
We empirically validated Ada-Instruct's efficacy across different applications, including code completion, mathematical reasoning, and commonsense reasoning.
- Score: 17.07852413707166
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generating diverse and sophisticated instructions for downstream tasks by
Large Language Models (LLMs) is pivotal for advancing the effect. Current
approaches leverage closed-source LLMs, employing in-context prompting for
instruction generation. However, in this paper, we found that in-context
prompting cannot generate complex instructions with length $\ge 100$ for tasks
like code completion.
To solve this problem, we introduce Ada-Instruct, an adaptive instruction
generator developed by fine-tuning open-source LLMs. Our pivotal finding
illustrates that fine-tuning open-source LLMs with a mere ten samples generates
long instructions that maintain distributional consistency for complex
reasoning tasks. We empirically validated Ada-Instruct's efficacy across
different applications, including code completion, mathematical reasoning, and
commonsense reasoning. The results underscore Ada-Instruct's superiority,
evidencing its improvements over its base models, current self-instruct
methods, and other state-of-the-art models.
Related papers
- AIR: Complex Instruction Generation via Automatic Iterative Refinement [29.639832268719363]
Current approaches to generating complex instructions are often irrelevant to the current instruction requirements.
We propose a novel automatic iterative refinement framework to generate complex instructions with constraints.
We construct the AIR-10K dataset with 10K complex instructions and demonstrate that instructions generated with our approach significantly improve the model's ability to follow complex instructions.
arXiv Detail & Related papers (2025-02-25T02:39:57Z) - Smaller Language Models Are Better Instruction Evolvers [10.587052565101844]
Small language models (SLMs) can synthesize more effective instructions than large language models (LLMs)
We propose Instruction Complex-Aware IFD (IC-IFD) to evaluate the effectiveness of instruction data more accurately.
arXiv Detail & Related papers (2024-12-15T16:07:48Z) - Evolutionary Contrastive Distillation for Language Model Alignment [35.94171633370035]
Evolutionary Contrastive Distillation (ECD) is a novel method for generating high-quality synthetic preference data.
Our method yields a 7B model that exceeds the complex instruction-following performance of current SOTA 7B models.
arXiv Detail & Related papers (2024-10-10T01:04:03Z) - TaCIE: Enhancing Instruction Comprehension in Large Language Models through Task-Centred Instruction Evolution [27.949846287419998]
TaCIE redefines instruction evolution from merely evolving seed instructions to a more dynamic and comprehensive combination of elements.
Applying TaCIE across multiple domains, LLMs fine-tuned with these evolved instructions have substantially outperformed those tuned with conventional methods.
arXiv Detail & Related papers (2024-09-18T10:06:28Z) - Controllable Navigation Instruction Generation with Chain of Thought Prompting [74.34604350917273]
We propose C-Instructor, which utilizes the chain-of-thought-style prompt for style-controllable and content-controllable instruction generation.
C-Instructor renders generated instructions more accessible to follow and offers greater controllability over the manipulation of landmark objects.
arXiv Detail & Related papers (2024-07-10T07:37:20Z) - One Prompt is not Enough: Automated Construction of a Mixture-of-Expert Prompts [110.94724216491753]
Large Language Models (LLMs) exhibit strong generalization capabilities when prompted with language instructions and in-context demos.
Various methods have been explored to automate the instruction design, but they restricted the searched prompt to one instruction.
We adopt the Mixture-of-Expert paradigm and divide the problem space into a set of sub-regions.
A two-phase process is developed to construct the specialized expert for each region.
A region-based joint search of an instruction per expert complements the demos assigned to it, yielding a synergistic effect.
arXiv Detail & Related papers (2024-06-28T23:05:08Z) - Chain-of-Instructions: Compositional Instruction Tuning on Large Language Models [15.444719480373001]
We propose a novel concept of compositional instructions called chain-of-instructions (CoI)
Unlike the conventional practice of solving single instruction tasks, our proposed method encourages a model to solve each subtask step by step until the final answer is reached.
CoI-tuning improves the model's ability to handle instructions composed of multiple subtasks as well as unseen composite tasks such as multilingual summarization.
arXiv Detail & Related papers (2024-02-18T10:10:40Z) - What Makes for Good Visual Instructions? Synthesizing Complex Visual Reasoning Instructions for Visual Instruction Tuning [111.01953096869947]
Visual instruction tuning is crucial for enhancing the zero-shot generalization capability of Multi-modal Large Language Models (MLLMs)
We develop a systematic approach to automatically create high-quality complex visual reasoning instructions.
Experimental results consistently demonstrate the enhanced performance of all compared MLLMs.
arXiv Detail & Related papers (2023-11-02T15:36:12Z) - Ensemble-Instruct: Generating Instruction-Tuning Data with a
Heterogeneous Mixture of LMs [23.38507910115345]
In-context learning (ICL) techniques can train strong conversational agents with only a small amount of human supervision.
Here we explore the application of such techniques to language models that are much smaller (around 10B--40B parameters) and have permissive licenses.
We find the Self-Instruct approach to be less effective at these sizes and propose new ICL methods that draw on two main ideas.
arXiv Detail & Related papers (2023-10-21T10:21:17Z) - From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning [63.63840740526497]
We investigate how instruction tuning adjusts pre-trained models with a focus on intrinsic changes.
The impact of instruction tuning is then studied by comparing the explanations derived from the pre-trained and instruction-tuned models.
Our findings reveal three significant impacts of instruction tuning.
arXiv Detail & Related papers (2023-09-30T21:16:05Z) - Can Large Language Models Understand Real-World Complex Instructions? [54.86632921036983]
Large language models (LLMs) can understand human instructions, but struggle with complex instructions.
Existing benchmarks are insufficient to assess LLMs' ability to understand complex instructions.
We propose CELLO, a benchmark for evaluating LLMs' ability to follow complex instructions systematically.
arXiv Detail & Related papers (2023-09-17T04:18:39Z) - Enhancing Large Language Models Against Inductive Instructions with
Dual-critique Prompting [55.15697111170836]
This paper reveals the behaviors of large language models (LLMs) towards textitinductive instructions and enhance their truthfulness and helpfulness accordingly.
After extensive human and automatic evaluations, we uncovered a universal vulnerability among LLMs in processing inductive instructions.
We identify that different inductive styles affect the models' ability to identify the same underlying errors, and the complexity of the underlying assumptions also influences the model's performance.
arXiv Detail & Related papers (2023-05-23T06:38:20Z) - WizardLM: Empowering Large Language Models to Follow Complex
Instructions [67.41048242052258]
We show an avenue for creating large amounts of instruction data with varying levels of complexity using LLM instead of humans.
We use our proposed Evol-Instruct to rewrite instructions step by step into more complex instructions.
Then, we mix all generated instruction data to fine-tune LLaMA.
arXiv Detail & Related papers (2023-04-24T16:31:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.