Meta Learning in Decentralized Neural Networks: Towards More General AI
- URL: http://arxiv.org/abs/2302.01020v1
- Date: Thu, 2 Feb 2023 11:15:07 GMT
- Title: Meta Learning in Decentralized Neural Networks: Towards More General AI
- Authors: Yuwei Sun
- Abstract summary: We aim to provide a fundamental understanding of learning to learn in the contents of Decentralized Neural Networks (Decentralized NNs)
We will present three different approaches to building such a decentralized learning system.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Meta-learning usually refers to a learning algorithm that learns from other
learning algorithms. The problem of uncertainty in the predictions of neural
networks shows that the world is only partially predictable and a learned
neural network cannot generalize to its ever-changing surrounding environments.
Therefore, the question is how a predictive model can represent multiple
predictions simultaneously. We aim to provide a fundamental understanding of
learning to learn in the contents of Decentralized Neural Networks
(Decentralized NNs) and we believe this is one of the most important questions
and prerequisites to building an autonomous intelligence machine. To this end,
we shall demonstrate several pieces of evidence for tackling the problems above
with Meta Learning in Decentralized NNs. In particular, we will present three
different approaches to building such a decentralized learning system: (1)
learning from many replica neural networks, (2) building the hierarchy of
neural networks for different functions, and (3) leveraging different modality
experts to learn cross-modal representations.
Related papers
- Collective variables of neural networks: empirical time evolution and scaling laws [0.535514140374842]
We show that certain measures on the spectrum of the empirical neural tangent kernel, specifically entropy and trace, yield insight into the representations learned by a neural network.
Results are demonstrated first on test cases before being shown on more complex networks, including transformers, auto-encoders, graph neural networks, and reinforcement learning studies.
arXiv Detail & Related papers (2024-10-09T21:37:14Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Reasoning Algorithmically in Graph Neural Networks [1.8130068086063336]
We aim to integrate the structured and rule-based reasoning of algorithms with adaptive learning capabilities of neural networks.
This dissertation provides theoretical and practical contributions to this area of research.
arXiv Detail & Related papers (2024-02-21T12:16:51Z) - ShadowNet for Data-Centric Quantum System Learning [188.683909185536]
We propose a data-centric learning paradigm combining the strength of neural-network protocols and classical shadows.
Capitalizing on the generalization power of neural networks, this paradigm can be trained offline and excel at predicting previously unseen systems.
We present the instantiation of our paradigm in quantum state tomography and direct fidelity estimation tasks and conduct numerical analysis up to 60 qubits.
arXiv Detail & Related papers (2023-08-22T09:11:53Z) - Meta Neural Coordination [0.0]
Meta-learning aims to develop algorithms that can learn from other learning algorithms to adapt to new and changing environments.
Uncertainty in the predictions of conventional deep neural networks highlights the partial predictability of the world.
We discuss the potential advancements required to build biologically-inspired machine intelligence.
arXiv Detail & Related papers (2023-05-20T06:06:44Z) - Synergistic information supports modality integration and flexible
learning in neural networks solving multiple tasks [107.8565143456161]
We investigate the information processing strategies adopted by simple artificial neural networks performing a variety of cognitive tasks.
Results show that synergy increases as neural networks learn multiple diverse tasks.
randomly turning off neurons during training through dropout increases network redundancy, corresponding to an increase in robustness.
arXiv Detail & Related papers (2022-10-06T15:36:27Z) - The least-control principle for learning at equilibrium [65.2998274413952]
We present a new principle for learning equilibrium recurrent neural networks, deep equilibrium models, or meta-learning.
Our results shed light on how the brain might learn and offer new ways of approaching a broad class of machine learning problems.
arXiv Detail & Related papers (2022-07-04T11:27:08Z) - Rank Diminishing in Deep Neural Networks [71.03777954670323]
Rank of neural networks measures information flowing across layers.
It is an instance of a key structural condition that applies across broad domains of machine learning.
For neural networks, however, the intrinsic mechanism that yields low-rank structures remains vague and unclear.
arXiv Detail & Related papers (2022-06-13T12:03:32Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - Introduction to Machine Learning for the Sciences [0.0]
The notes start with an exposition of machine learning methods without neural networks, such as principle component analysis, t-SNE, and linear regression.
We continue with an introduction to both basic and advanced neural network structures such as conventional neural networks, (variational) autoencoders, generative adversarial networks, restricted Boltzmann machines, and recurrent neural networks.
arXiv Detail & Related papers (2021-02-08T16:25:46Z) - Locality Guided Neural Networks for Explainable Artificial Intelligence [12.435539489388708]
We propose a novel algorithm for back propagation, called Locality Guided Neural Network(LGNN)
LGNN preserves locality between neighbouring neurons within each layer of a deep network.
In our experiments, we train various VGG and Wide ResNet (WRN) networks for image classification on CIFAR100.
arXiv Detail & Related papers (2020-07-12T23:45:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.