Modularizing Deep Learning via Pairwise Learning With Kernels
- URL: http://arxiv.org/abs/2005.05541v2
- Date: Thu, 10 Sep 2020 18:09:47 GMT
- Title: Modularizing Deep Learning via Pairwise Learning With Kernels
- Authors: Shiyu Duan, Shujian Yu, Jose Principe
- Abstract summary: We present an alternative view on finitely wide, fully trainable deep computation neural networks as stacked linear models in feature spaces.
We then propose a provably optimal modular learning framework for classification that does not require between- module backpropagation.
- Score: 12.051746916737343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: By redefining the conventional notions of layers, we present an alternative
view on finitely wide, fully trainable deep neural networks as stacked linear
models in feature spaces, leading to a kernel machine interpretation. Based on
this construction, we then propose a provably optimal modular learning
framework for classification that does not require between-module
backpropagation. This modular approach brings new insights into the label
requirement of deep learning: It leverages only implicit pairwise labels (weak
supervision) when learning the hidden modules. When training the output module,
on the other hand, it requires full supervision but achieves high label
efficiency, needing as few as 10 randomly selected labeled examples (one from
each class) to achieve 94.88% accuracy on CIFAR-10 using a ResNet-18 backbone.
Moreover, modular training enables fully modularized deep learning workflows,
which then simplify the design and implementation of pipelines and improve the
maintainability and reusability of models. To showcase the advantages of such a
modularized workflow, we describe a simple yet reliable method for estimating
reusability of pre-trained modules as well as task transferability in a
transfer learning setting. At practically no computation overhead, it precisely
described the task space structure of 15 binary classification tasks from
CIFAR-10.
Related papers
- Train Faster, Perform Better: Modular Adaptive Training in Over-Parameterized Models [31.960749305728488]
We introduce a novel concept dubbed modular neural tangent kernel (mNTK)
We show that the quality of a module's learning is tightly associated with its mNTK's principal eigenvalue $lambda_max$.
We propose a novel training strategy termed Modular Adaptive Training (MAT) to update those modules with their $lambda_max$ exceeding a dynamic threshold.
arXiv Detail & Related papers (2024-05-13T07:46:48Z) - Modular Deep Learning [120.36599591042908]
Transfer learning has recently become the dominant paradigm of machine learning.
It remains unclear how to develop models that specialise towards multiple tasks without incurring negative interference.
Modular deep learning has emerged as a promising solution to these challenges.
arXiv Detail & Related papers (2023-02-22T18:11:25Z) - Unifying Synergies between Self-supervised Learning and Dynamic
Computation [53.66628188936682]
We present a novel perspective on the interplay between SSL and DC paradigms.
We show that it is feasible to simultaneously learn a dense and gated sub-network from scratch in a SSL setting.
The co-evolution during pre-training of both dense and gated encoder offers a good accuracy-efficiency trade-off.
arXiv Detail & Related papers (2023-01-22T17:12:58Z) - Local Learning with Neuron Groups [15.578925277062657]
Local learning is an approach to model-parallelism that removes the standard end-to-end learning setup.
We study how local learning can be applied at the level of splitting layers or modules into sub-components.
arXiv Detail & Related papers (2023-01-18T16:25:10Z) - Padding Module: Learning the Padding in Deep Neural Networks [4.769747792846005]
This paper proposes a trainable Padding Module that can be placed in a deep learning model.
The Padding Module can optimize itself without requiring or influencing the model's entire loss function.
Experiments have shown that the proposed Padding Module outperforms the state-of-the-art competitors and the baseline methods.
arXiv Detail & Related papers (2023-01-11T18:03:57Z) - Learning Label Modular Prompts for Text Classification in the Wild [56.66187728534808]
We propose text classification in-the-wild, which introduces different non-stationary training/testing stages.
Decomposing a complex task into modular components can enable robust generalisation under such non-stationary environment.
We propose MODULARPROMPT, a label-modular prompt tuning framework for text classification tasks.
arXiv Detail & Related papers (2022-11-30T16:26:38Z) - Neural Network Module Decomposition and Recomposition [35.21448933547118]
We propose a modularization method that decomposes a deep neural network (DNN) into small modules from a functionality perspective.
We demonstrate that the proposed method can decompose and recompose DNNs with high compression ratio and high accuracy.
arXiv Detail & Related papers (2021-12-25T08:36:47Z) - Weakly Supervised Semantic Segmentation via Alternative Self-Dual
Teaching [82.71578668091914]
This paper establishes a compact learning framework that embeds the classification and mask-refinement components into a unified deep model.
We propose a novel alternative self-dual teaching (ASDT) mechanism to encourage high-quality knowledge interaction.
arXiv Detail & Related papers (2021-12-17T11:56:56Z) - Knodle: Modular Weakly Supervised Learning with PyTorch [5.874587993411972]
Knodle is a software framework for separating weak data annotations, powerful deep learning models, and methods for improving weakly supervised training.
This modularization gives the training process access to fine-grained information such as data set characteristics, matches of rules, or elements of the deep learning model ultimately used for prediction.
arXiv Detail & Related papers (2021-04-23T12:33:25Z) - Neural Function Modules with Sparse Arguments: A Dynamic Approach to
Integrating Information across Layers [84.57980167400513]
Neural Function Modules (NFM) aims to introduce the same structural capability into deep learning.
Most of the work in the context of feed-forward networks combining top-down and bottom-up feedback is limited to classification problems.
The key contribution of our work is to combine attention, sparsity, top-down and bottom-up feedback, in a flexible algorithm.
arXiv Detail & Related papers (2020-10-15T20:43:17Z) - Fast Few-Shot Classification by Few-Iteration Meta-Learning [173.32497326674775]
We introduce a fast optimization-based meta-learning method for few-shot classification.
Our strategy enables important aspects of the base learner objective to be learned during meta-training.
We perform a comprehensive experimental analysis, demonstrating the speed and effectiveness of our approach.
arXiv Detail & Related papers (2020-10-01T15:59:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.