Curriculum-Meta Learning for Order-Robust Continual Relation Extraction
- URL: http://arxiv.org/abs/2101.01926v3
- Date: Fri, 8 Jan 2021 10:06:40 GMT
- Title: Curriculum-Meta Learning for Order-Robust Continual Relation Extraction
- Authors: Tongtong Wu, Xuekai Li, Yuan-Fang Li, Reza Haffari, Guilin Qi, Yujin
Zhu and Guoqiang Xu
- Abstract summary: We propose a novel curriculum-meta learning method to tackle the challenges of continual relation extraction.
We combine meta learning and curriculum learning to quickly adapt model parameters to a new task.
We present novel difficulty-based metrics to quantitatively measure the extent of order-sensitivity of a given model.
- Score: 12.494209368988253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continual relation extraction is an important task that focuses on extracting
new facts incrementally from unstructured text. Given the sequential arrival
order of the relations, this task is prone to two serious challenges, namely
catastrophic forgetting and order-sensitivity. We propose a novel
curriculum-meta learning method to tackle the above two challenges in continual
relation extraction. We combine meta learning and curriculum learning to
quickly adapt model parameters to a new task and to reduce interference of
previously seen tasks on the current task. We design a novel relation
representation learning method through the distribution of domain and range
types of relations. Such representations are utilized to quantify the
difficulty of tasks for the construction of curricula. Moreover, we also
present novel difficulty-based metrics to quantitatively measure the extent of
order-sensitivity of a given model, suggesting new ways to evaluate model
robustness. Our comprehensive experiments on three benchmark datasets show that
our proposed method outperforms the state-of-the-art techniques. The code is
available at the anonymous GitHub repository:
https://github.com/wutong8023/AAAI_CML.
Related papers
- Learning Task Representations from In-Context Learning [73.72066284711462]
Large language models (LLMs) have demonstrated remarkable proficiency in in-context learning.
We introduce an automated formulation for encoding task information in ICL prompts as a function of attention heads.
We show that our method's effectiveness stems from aligning the distribution of the last hidden state with that of an optimally performing in-context-learned model.
arXiv Detail & Related papers (2025-02-08T00:16:44Z) - A Tensor Low-Rank Approximation for Value Functions in Multi-Task Reinforcement Learning [10.359616364592073]
In pursuit of reinforcement learning systems that could train in physical environments, we investigate multi-task approaches.
A low-rank structure enforces the notion of similarity, without the need to explicitly prescribe which tasks are similar.
The efficiency of our low-rank tensor approach to multi-task learning is demonstrated in two numerical experiments.
arXiv Detail & Related papers (2025-01-17T20:07:11Z) - Data-CUBE: Data Curriculum for Instruction-based Sentence Representation
Learning [85.66907881270785]
We propose a data curriculum method, namely Data-CUBE, that arranges the orders of all the multi-task data for training.
In the task level, we aim to find the optimal task order to minimize the total cross-task interference risk.
In the instance level, we measure the difficulty of all instances per task, then divide them into the easy-to-difficult mini-batches for training.
arXiv Detail & Related papers (2024-01-07T18:12:20Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Continual Few-shot Relation Learning via Embedding Space Regularization
and Data Augmentation [4.111899441919165]
It is necessary for the model to learn novel relational patterns with very few labeled data while avoiding catastrophic forgetting of previous task knowledge.
We propose a novel method based on embedding space regularization and data augmentation.
Our method generalizes to new few-shot tasks and avoids catastrophic forgetting of previous tasks by enforcing extra constraints on the relational embeddings and by adding extra relevant data in a self-supervised manner.
arXiv Detail & Related papers (2022-03-04T05:19:09Z) - Relational Experience Replay: Continual Learning by Adaptively Tuning
Task-wise Relationship [54.73817402934303]
We propose Experience Continual Replay (ERR), a bi-level learning framework to adaptively tune task-wise to achieve a better stability plasticity' tradeoff.
ERR can consistently improve the performance of all baselines and surpass current state-of-the-art methods.
arXiv Detail & Related papers (2021-12-31T12:05:22Z) - Exploring Task Difficulty for Few-Shot Relation Extraction [22.585574542329677]
Few-shot relation extraction (FSRE) focuses on recognizing novel relations by learning with merely a handful of annotated instances.
We introduce a novel approach based on contrastive learning that learns better representations by exploiting relation label information.
arXiv Detail & Related papers (2021-09-12T09:40:33Z) - Learning Invariant Representation for Continual Learning [5.979373021392084]
A key challenge in Continual learning is catastrophically forgetting previously learned tasks when the agent faces a new one.
We propose a new pseudo-rehearsal-based method, named learning Invariant Representation for Continual Learning (IRCL)
Disentangling the shared invariant representation helps to learn continually a sequence of tasks, while being more robust to forgetting and having better knowledge transfer.
arXiv Detail & Related papers (2021-01-15T15:12:51Z) - Meta-Reinforcement Learning Robust to Distributional Shift via Model
Identification and Experience Relabeling [126.69933134648541]
We present a meta-reinforcement learning algorithm that is both efficient and extrapolates well when faced with out-of-distribution tasks at test time.
Our method is based on a simple insight: we recognize that dynamics models can be adapted efficiently and consistently with off-policy data.
arXiv Detail & Related papers (2020-06-12T13:34:46Z) - Meta Cyclical Annealing Schedule: A Simple Approach to Avoiding
Meta-Amortization Error [50.83356836818667]
We develop a novel meta-regularization objective using it cyclical annealing schedule and it maximum mean discrepancy (MMD) criterion.
The experimental results show that our approach substantially outperforms standard meta-learning algorithms.
arXiv Detail & Related papers (2020-03-04T04:43:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.