Exploring the Intersection between Neural Architecture Search and
Continual Learning
- URL: http://arxiv.org/abs/2206.05625v2
- Date: Thu, 15 Jun 2023 17:04:02 GMT
- Title: Exploring the Intersection between Neural Architecture Search and
Continual Learning
- Authors: Mohamed Shahawy, Elhadj Benkhelifa, David White
- Abstract summary: Continual adaptiveness and automation of neural networks is of paramount importance to several domains.
This study conducts the first extensive review on the intersection between Neural Architecture Search (NAS) and Continual Learning (CL)
- Score: 0.2148535041822524
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite the significant advances achieved in Artificial Neural Networks
(ANNs), their design process remains notoriously tedious, depending primarily
on intuition, experience and trial-and-error. This human-dependent process is
often time-consuming and prone to errors. Furthermore, the models are generally
bound to their training contexts, with no considerations to their surrounding
environments. Continual adaptiveness and automation of neural networks is of
paramount importance to several domains where model accessibility is limited
after deployment (e.g IoT devices, self-driving vehicles, etc.). Additionally,
even accessible models require frequent maintenance post-deployment to overcome
issues such as Concept/Data Drift, which can be cumbersome and restrictive. By
leveraging and combining approaches from Neural Architecture Search (NAS) and
Continual Learning (CL), more robust and adaptive agents can be developed. This
study conducts the first extensive review on the intersection between NAS and
CL, formalizing the prospective Continually-Adaptive Neural Networks (CANNs)
paradigm and outlining research directions for lifelong autonomous ANNs.
Related papers
- Continual Learning with Hebbian Plasticity in Sparse and Predictive Coding Networks: A Survey and Perspective [1.3986052523534573]
An emerging class of neuromorphic continual learning systems must learn to integrate new information on the fly.
This survey covers a number of recent works in the field of neuromorphic continual learning based on state-of-the-art Sparse and Predictive Coding technology.
It is hoped that this survey will contribute towards future research in the field of neuromorphic continual learning.
arXiv Detail & Related papers (2024-07-24T14:20:59Z) - NNsight and NDIF: Democratizing Access to Open-Weight Foundation Model Internals [58.83169560132308]
We introduce NNsight and NDIF, technologies that work in tandem to enable scientific study of the representations and computations learned by very large neural networks.
arXiv Detail & Related papers (2024-07-18T17:59:01Z) - The Dynamic Net Architecture: Learning Robust and Holistic Visual Representations Through Self-Organizing Networks [3.9848584845601014]
We present a novel intelligent-system architecture called "Dynamic Net Architecture" (DNA)
DNA relies on recurrence-stabilized networks and discuss it in application to vision.
arXiv Detail & Related papers (2024-07-08T06:22:10Z) - ELiSe: Efficient Learning of Sequences in Structured Recurrent Networks [1.5931140598271163]
We build a model for efficient learning sequences using only local always-on and phase-free plasticity.
We showcase the capabilities of ELiSe in a mock-up of birdsong learning, and demonstrate its flexibility with respect to parametrization.
arXiv Detail & Related papers (2024-02-26T17:30:34Z) - Neuro-mimetic Task-free Unsupervised Online Learning with Continual
Self-Organizing Maps [56.827895559823126]
Self-organizing map (SOM) is a neural model often used in clustering and dimensionality reduction.
We propose a generalization of the SOM, the continual SOM, which is capable of online unsupervised learning under a low memory budget.
Our results, on benchmarks including MNIST, Kuzushiji-MNIST, and Fashion-MNIST, show almost a two times increase in accuracy.
arXiv Detail & Related papers (2024-02-19T19:11:22Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Efficient Neural Architecture Search with Performance Prediction [0.0]
We use a neural architecture search to find the best network architecture for the task at hand.
Existing NAS algorithms generally evaluate the fitness of a new architecture by fully training from scratch.
An end-to-end offline performance predictor is proposed to accelerate the evaluation of sampled architectures.
arXiv Detail & Related papers (2021-08-04T05:44:16Z) - Dependable Neural Networks Through Redundancy, A Comparison of Redundant
Architectures [0.0]
We discuss the issue of dependability in general implementation terms before examining lockstep solutions.
We intuit that it is not necessarily a given that two similar neural networks generate results at precisely the same time and that synchronization between the platforms will be required.
arXiv Detail & Related papers (2021-07-30T12:17:36Z) - Causal Navigation by Continuous-time Neural Networks [108.84958284162857]
We propose a theoretical and experimental framework for learning causal representations using continuous-time neural networks.
We evaluate our method in the context of visual-control learning of drones over a series of complex tasks.
arXiv Detail & Related papers (2021-06-15T17:45:32Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.