KRNet: Towards Efficient Knowledge Replay
- URL: http://arxiv.org/abs/2205.11126v1
- Date: Mon, 23 May 2022 08:34:17 GMT
- Title: KRNet: Towards Efficient Knowledge Replay
- Authors: Yingying Zhang, Qiaoyong Zhong, Di Xie, Shiliang Pu
- Abstract summary: A knowledge replay technique has been widely used in many tasks such as continual learning and continuous domain adaptation.
We propose a novel and efficient knowledge recording network (KRNet) which directly maps an arbitrary sample identity number to the corresponding datum.
Our KRNet requires significantly less storage cost for the latent codes and can be trained without the encoder sub-network.
- Score: 50.315451023983805
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The knowledge replay technique has been widely used in many tasks such as
continual learning and continuous domain adaptation. The key lies in how to
effectively encode the knowledge extracted from previous data and replay them
during current training procedure. A simple yet effective model to achieve
knowledge replay is autoencoder. However, the number of stored latent codes in
autoencoder increases linearly with the scale of data and the trained encoder
is redundant for the replaying stage. In this paper, we propose a novel and
efficient knowledge recording network (KRNet) which directly maps an arbitrary
sample identity number to the corresponding datum. Compared with autoencoder,
our KRNet requires significantly ($400\times$) less storage cost for the latent
codes and can be trained without the encoder sub-network. Extensive experiments
validate the efficiency of KRNet, and as a showcase, it is successfully applied
in the task of continual learning.
Related papers
- Regress Before Construct: Regress Autoencoder for Point Cloud
Self-supervised Learning [18.10704604275133]
Masked Autoencoders (MAE) have demonstrated promising performance in self-supervised learning for 2D and 3D computer vision.
We propose Point Regress AutoEncoder (Point-RAE), a new scheme for regressive autoencoders for point cloud self-supervised learning.
Our approach is efficient during pre-training and generalizes well on various downstream tasks.
arXiv Detail & Related papers (2023-09-25T17:23:33Z) - TransCoder: Towards Unified Transferable Code Representation Learning Inspired by Human Skills [31.75121546422898]
We present TransCoder, a unified Transferable fine-tuning strategy for Code representation learning.
We employ a tunable prefix encoder as the meta-learner to capture cross-task and cross-language transferable knowledge.
Our method can lead to superior performance on various code-related tasks and encourage mutual reinforcement.
arXiv Detail & Related papers (2023-05-23T06:59:22Z) - PEOPL: Characterizing Privately Encoded Open Datasets with Public Labels [59.66777287810985]
We introduce information-theoretic scores for privacy and utility, which quantify the average performance of an unfaithful user.
We then theoretically characterize primitives in building families of encoding schemes that motivate the use of random deep neural networks.
arXiv Detail & Related papers (2023-03-31T18:03:53Z) - Discrete Key-Value Bottleneck [95.61236311369821]
Deep neural networks perform well on classification tasks where data streams are i.i.d. and labeled data is abundant.
One powerful approach that has addressed this challenge involves pre-training of large encoders on volumes of readily available data, followed by task-specific tuning.
Given a new task, however, updating the weights of these encoders is challenging as a large number of weights needs to be fine-tuned, and as a result, they forget information about the previous tasks.
We propose a model architecture to address this issue, building upon a discrete bottleneck containing pairs of separate and learnable key-value codes.
arXiv Detail & Related papers (2022-07-22T17:52:30Z) - Pre-Training Transformer Decoder for End-to-End ASR Model with Unpaired
Speech Data [145.95460945321253]
We introduce two pre-training tasks for the encoder-decoder network using acoustic units, i.e., pseudo codes.
The proposed Speech2C can relatively reduce the word error rate (WER) by 19.2% over the method without decoder pre-training.
arXiv Detail & Related papers (2022-03-31T15:33:56Z) - A manifold learning perspective on representation learning: Learning
decoder and representations without an encoder [0.0]
Autoencoders are commonly used in representation learning.
Inspired by manifold learning, we show that the decoder can be trained on its own by learning the representations of the training samples.
Our approach of training the decoder alone facilitates representation learning even on small data sets.
arXiv Detail & Related papers (2021-08-31T15:08:50Z) - Recurrent autoencoder with sequence-aware encoding [0.0]
We propose an autoencoder architecture with sequence-aware encoding, which employs 1D convolutional layer to improve its performance.
We prove that the proposed solution dominates over the standard RAE, and the training process is order of magnitude faster.
arXiv Detail & Related papers (2020-09-15T20:51:20Z) - Decoupling Representation Learning from Reinforcement Learning [89.82834016009461]
We introduce an unsupervised learning task called Augmented Temporal Contrast (ATC)
ATC trains a convolutional encoder to associate pairs of observations separated by a short time difference.
In online RL experiments, we show that training the encoder exclusively using ATC matches or outperforms end-to-end RL.
arXiv Detail & Related papers (2020-09-14T19:11:13Z) - Hierarchical Memory Decoding for Video Captioning [43.51506421744577]
Memory network (MemNet) has the advantage of storing long-term information.
MemNet has not been well exploited for video captioning.
In this paper, we devise a novel memory decoder for video captioning.
arXiv Detail & Related papers (2020-02-27T02:48:10Z) - Encoding-based Memory Modules for Recurrent Neural Networks [79.42778415729475]
We study the memorization subtask from the point of view of the design and training of recurrent neural networks.
We propose a new model, the Linear Memory Network, which features an encoding-based memorization component built with a linear autoencoder for sequences.
arXiv Detail & Related papers (2020-01-31T11:14:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.