Teaching According to Talents! Instruction Tuning LLMs with Competence-Aware Curriculum Learning
- URL: http://arxiv.org/abs/2509.13790v2
- Date: Mon, 03 Nov 2025 09:06:01 GMT
- Title: Teaching According to Talents! Instruction Tuning LLMs with Competence-Aware Curriculum Learning
- Authors: Yangning Li, Tingwei Lu, Yinghui Li, Yankai Chen, Wei-Chieh Huang, Wenhao Jiang, Hui Wang, Hai-Tao Zheng, Philip S. Yu,
- Abstract summary: This paper presents a Competence-Aware Multi-Perspective cUrriculum inStruction tuning framework termed CAMPUS.<n> CAMPUS offers several advantages: Dynamic selection for sub-curriculum, competency-aware adjustment to the curriculum schedule, and multiple difficulty-based scheduling.
- Score: 64.92967672226534
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Efficient instruction tuning aims to enhance the ultimate performance of large language models (LLMs) trained on a given instruction dataset. Curriculum learning as a typical data organization strategy has shown preliminary effectiveness in instruction tuning. However, current curriculum tuning methods suffer from the curriculum rigidity, since they rely solely on static heuristic difficulty metrics. These methods fail to adapt to the evolving capabilities of models during training, resulting in a fixed and potentially sub-optimal learning trajectory. To address the issue, Competence-Aware Multi-Perspective cUrriculum inStruction tuning framework termed CAMPUS is proposed. CAMPUS offers several advantages: (1) Dynamic selection for sub-curriculum. (2) Competency-aware adjustment to the curriculum schedule. (3) Multiple difficulty-based scheduling. Extensive experiments prove the superior performance of CAMPUS, compared to other state-of-the-art baselines for efficient instruction tuning.
Related papers
- Bandit Guided Submodular Curriculum for Adaptive Subset Selection [12.516248058768264]
Traditional curriculum learning proceeds from easy to hard samples, yet defining a reliable notion of difficulty remains elusive.<n>We reinterpret adaptive subset selection and formulate it as a multi-armed bandit problem, where each arm corresponds to a submodular function guiding sample selection.<n>We introduce ONLINESUBMOD, a novel online greedy policy that optimize a utility-driven reward and provably achieves no-regret performance under various sampling regimes.
arXiv Detail & Related papers (2025-11-28T07:31:53Z) - CLASS-IT: Conversational and Lecture-Aligned Small-Scale Instruction Tuning for BabyLMs [81.79228604962687]
This work investigates whether small-scale LMs can benefit from instruction tuning.<n>We compare conversational and question-answering instruction tuning datasets, applied either in a merged or sequential curriculum.<n>Results show that instruction tuning yields small but consistent gains in fine-tuning scenarios, with sequential curricula outperforming merged data.<n>However, improvements do not consistently transfer to zero-shot tasks, suggesting a trade-off between interaction-focused adaptation and broad linguistic generalization.
arXiv Detail & Related papers (2025-10-29T10:36:39Z) - Your Pretrained Model Tells the Difficulty Itself: A Self-Adaptive Curriculum Learning Paradigm for Natural Language Understanding [53.63482987410292]
We present a self-adaptive curriculum learning paradigm that prioritizes fine-tuning examples based on difficulty scores predicted by pre-trained language models.<n>We evaluate our method on four natural language understanding (NLU) datasets covering both binary and multi-class classification tasks.
arXiv Detail & Related papers (2025-07-13T19:36:17Z) - Beyond Random Sampling: Efficient Language Model Pretraining via Curriculum Learning [23.900888224619]
We show that curriculum learning consistently improves convergence in early and mid-training phases.<n>We identify compression ratio, lexical diversity, and readability as effective difficulty signals across settings.
arXiv Detail & Related papers (2025-06-12T21:06:57Z) - Self-Evolving Curriculum for LLM Reasoning [96.10277986436172]
Self-Evolving Curriculum (SEC) is an automatic curriculum learning method that learns a curriculum policy concurrently with the RL fine-tuning process.<n>Our experiments demonstrate that SEC significantly improves models' reasoning capabilities, enabling better generalization to harder, out-of-distribution test problems.
arXiv Detail & Related papers (2025-05-20T23:17:15Z) - RAISE: Reinforced Adaptive Instruction Selection For Large Language Models [48.63476198469349]
We propose a task-objective-driven instruction selection framework RAISE(Reinforced Adaptive Instruction SElection)<n> RAISE incorporates the entire instruction fine-tuning process into optimization, selecting instructions at each step based on the expected impact of each instruction on model performance improvement.<n>Experiments and result analysis prove the superiority of our method compared with other instruction selection methods.
arXiv Detail & Related papers (2025-04-09T21:17:52Z) - MLAN: Language-Based Instruction Tuning Preserves and Transfers Knowledge in Multimodal Language Models [79.0546136194314]
We present a novel visual instruction tuning strategy to improve the zero-shot task generalization of multimodal large language models.<n>We find that simply increasing sufficiently diverse text-only data enables transfer of instruction following ability and domain knowledge across modalities while being more efficient than the vision-language approach.
arXiv Detail & Related papers (2024-11-15T20:09:59Z) - SwitchCIT: Switching for Continual Instruction Tuning [14.085371250265224]
Large language models (LLMs) and multimodal models (MMs) have exhibited impressive capabilities in various domains.<n>Continual instruction tuning is crucial to adapt a large model to evolving tasks and domains.<n>This work addresses the catastrophic forgetting in continual instruction learning through a mechanism for routing computations to parameter-efficient tuned models.
arXiv Detail & Related papers (2024-07-16T14:37:33Z) - One-Shot Learning as Instruction Data Prospector for Large Language Models [108.81681547472138]
textscNuggets uses one-shot learning to select high-quality instruction data from extensive datasets.
We show that instruction tuning with the top 1% of examples curated by textscNuggets substantially outperforms conventional methods employing the entire dataset.
arXiv Detail & Related papers (2023-12-16T03:33:12Z) - When Do Curricula Work? [26.072472732516335]
ordered learning has been suggested as improvements to the standard i.i.d. training.
We conduct experiments over thousands of orderings spanning three kinds of learning: curriculum, anti-curriculum, and random-curriculum.
We find that curricula have only marginal benefits, and that randomly ordered samples perform as well or better than curricula and anti-curricula.
arXiv Detail & Related papers (2020-12-05T19:41:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.