MOLE: MOdular Learning FramEwork via Mutual Information Maximization
- URL: http://arxiv.org/abs/2308.07772v1
- Date: Tue, 15 Aug 2023 13:48:16 GMT
- Title: MOLE: MOdular Learning FramEwork via Mutual Information Maximization
- Authors: Tianchao Li and Yulong Pei
- Abstract summary: This paper introduces an asynchronous and local learning framework for neural networks, named Modular Learning Framework (MOLE)
MOLE modularizes neural networks by layers, defines the training objective via mutual information for each module, and sequentially trains each module by mutual information.
In particular, this framework is capable of solving both graph- and node-level tasks for graph-type data.
- Score: 3.8399484206282146
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper is to introduce an asynchronous and local learning framework for
neural networks, named Modular Learning Framework (MOLE). This framework
modularizes neural networks by layers, defines the training objective via
mutual information for each module, and sequentially trains each module by
mutual information maximization. MOLE makes the training become local
optimization with gradient-isolated across modules, and this scheme is more
biologically plausible than BP. We run experiments on vector-, grid- and
graph-type data. In particular, this framework is capable of solving both
graph- and node-level tasks for graph-type data. Therefore, MOLE has been
experimentally proven to be universally applicable to different types of data.
Related papers
- Train Faster, Perform Better: Modular Adaptive Training in Over-Parameterized Models [31.960749305728488]
We introduce a novel concept dubbed modular neural tangent kernel (mNTK)
We show that the quality of a module's learning is tightly associated with its mNTK's principal eigenvalue $lambda_max$.
We propose a novel training strategy termed Modular Adaptive Training (MAT) to update those modules with their $lambda_max$ exceeding a dynamic threshold.
arXiv Detail & Related papers (2024-05-13T07:46:48Z) - R-Cut: Enhancing Explainability in Vision Transformers with Relationship
Weighted Out and Cut [14.382326829600283]
We introduce two modules: the Relationship Weighted Out" and the Cut" modules.
The Cut" module performs fine-grained feature decomposition, taking into account factors such as position, texture, and color.
We validate our method with extensive qualitative and quantitative experiments on the ImageNet dataset.
arXiv Detail & Related papers (2023-07-18T08:03:51Z) - Decouple Graph Neural Networks: Train Multiple Simple GNNs Simultaneously Instead of One [60.5818387068983]
Graph neural networks (GNN) suffer from severe inefficiency.
We propose to decouple a multi-layer GNN as multiple simple modules for more efficient training.
We show that the proposed framework is highly efficient with reasonable performance.
arXiv Detail & Related papers (2023-04-20T07:21:32Z) - Modular Deep Learning [120.36599591042908]
Transfer learning has recently become the dominant paradigm of machine learning.
It remains unclear how to develop models that specialise towards multiple tasks without incurring negative interference.
Modular deep learning has emerged as a promising solution to these challenges.
arXiv Detail & Related papers (2023-02-22T18:11:25Z) - Neural Attentive Circuits [93.95502541529115]
We introduce a general purpose, yet modular neural architecture called Neural Attentive Circuits (NACs)
NACs learn the parameterization and a sparse connectivity of neural modules without using domain knowledge.
NACs achieve an 8x speedup at inference time while losing less than 3% performance.
arXiv Detail & Related papers (2022-10-14T18:00:07Z) - ModLaNets: Learning Generalisable Dynamics via Modularity and Physical
Inductive Bias [14.474273671369584]
We propose a structural neural network framework with modularity and physical inductive bias.
This framework models the energy of each element using modularity and then construct the target dynamical system via Lagrangian mechanics.
We examine our framework for modelling double-pendulum or three-body systems with small training datasets.
arXiv Detail & Related papers (2022-06-24T14:54:25Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Neural Function Modules with Sparse Arguments: A Dynamic Approach to
Integrating Information across Layers [84.57980167400513]
Neural Function Modules (NFM) aims to introduce the same structural capability into deep learning.
Most of the work in the context of feed-forward networks combining top-down and bottom-up feedback is limited to classification problems.
The key contribution of our work is to combine attention, sparsity, top-down and bottom-up feedback, in a flexible algorithm.
arXiv Detail & Related papers (2020-10-15T20:43:17Z) - S2RMs: Spatially Structured Recurrent Modules [105.0377129434636]
We take a step towards exploiting dynamic structure that are capable of simultaneously exploiting both modular andtemporal structures.
We find our models to be robust to the number of available views and better capable of generalization to novel tasks without additional training.
arXiv Detail & Related papers (2020-07-13T17:44:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.