Continual Learning with Deep Learning Methods in an Application-Oriented
Context
- URL: http://arxiv.org/abs/2207.06233v1
- Date: Tue, 12 Jul 2022 10:13:33 GMT
- Title: Continual Learning with Deep Learning Methods in an Application-Oriented
Context
- Authors: Benedikt Pf\"ulb
- Abstract summary: An important research area of Artificial Intelligence (AI) deals with the automatic derivation of knowledge from data.
One type of machine learning algorithms that can be categorized as "deep learning" model is referred to as Deep Neural Networks (DNNs)
DNNs are affected by a problem that prevents new knowledge from being added to an existing base.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Abstract knowledge is deeply grounded in many computer-based applications. An
important research area of Artificial Intelligence (AI) deals with the
automatic derivation of knowledge from data. Machine learning offers the
according algorithms. One area of research focuses on the development of
biologically inspired learning algorithms. The respective machine learning
methods are based on neurological concepts so that they can systematically
derive knowledge from data and store it. One type of machine learning
algorithms that can be categorized as "deep learning" model is referred to as
Deep Neural Networks (DNNs). DNNs consist of multiple artificial neurons
arranged in layers that are trained by using the backpropagation algorithm.
These deep learning methods exhibit amazing capabilities for inferring and
storing complex knowledge from high-dimensional data. However, DNNs are
affected by a problem that prevents new knowledge from being added to an
existing base. The ability to continuously accumulate knowledge is an important
factor that contributed to evolution and is therefore a prerequisite for the
development of strong AIs. The so-called "catastrophic forgetting" (CF) effect
causes DNNs to immediately loose already derived knowledge after a few training
iterations on a new data distribution. Only an energetically expensive
retraining with the joint data distribution of past and new data enables the
abstraction of the entire new set of knowledge. In order to counteract the
effect, various techniques have been and are still being developed with the
goal to mitigate or even solve the CF problem. These published CF avoidance
studies usually imply the effectiveness of their approaches for various
continual learning tasks. This dissertation is set in the context of continual
machine learning with deep learning methods. The first part deals with the
development of an ...
Related papers
- A Methodology-Oriented Study of Catastrophic Forgetting in Incremental Deep Neural Networks [4.844142799909677]
The ability of learning throughout their lifespan is referred as continuous learning which is using neurocognition mechanism.
The main challenge in artificial intelligence is that the incremental learning of the autonomous agent when new data confronted.
Here we are comparing three types of incremental learning methods: (1) Exemplar based methods, (2) Memory based methods, and (3) Network based method.
arXiv Detail & Related papers (2024-05-11T05:10:07Z) - Hebbian Learning based Orthogonal Projection for Continual Learning of
Spiking Neural Networks [74.3099028063756]
We develop a new method with neuronal operations based on lateral connections and Hebbian learning.
We show that Hebbian and anti-Hebbian learning on recurrent lateral connections can effectively extract the principal subspace of neural activities.
Our method consistently solves for spiking neural networks with nearly zero forgetting.
arXiv Detail & Related papers (2024-02-19T09:29:37Z) - Breaking the Curse of Dimensionality in Deep Neural Networks by Learning
Invariant Representations [1.9580473532948401]
This thesis explores the theoretical foundations of deep learning by studying the relationship between the architecture of these models and the inherent structures found within the data they process.
We ask What drives the efficacy of deep learning algorithms and allows them to beat the so-called curse of dimensionality.
Our methodology takes an empirical approach to deep learning, combining experimental studies with physics-inspired toy models.
arXiv Detail & Related papers (2023-10-24T19:50:41Z) - How Efficient Are Today's Continual Learning Algorithms? [31.120016345185217]
Supervised continual learning involves updating a deep neural network (DNN) from an ever-growing stream of labeled data.
One of the major motivations behind continual learning is being able to efficiently update a network with new information, rather than retraining from scratch on the training dataset as it grows over time.
Here, we study recent methods for incremental class learning and illustrate that many are highly inefficient in terms of compute, memory, and storage.
arXiv Detail & Related papers (2023-03-29T18:52:10Z) - Neuro-Symbolic Learning of Answer Set Programs from Raw Data [54.56905063752427]
Neuro-Symbolic AI aims to combine interpretability of symbolic techniques with the ability of deep learning to learn from raw data.
We introduce Neuro-Symbolic Inductive Learner (NSIL), an approach that trains a general neural network to extract latent concepts from raw data.
NSIL learns expressive knowledge, solves computationally complex problems, and achieves state-of-the-art performance in terms of accuracy and data efficiency.
arXiv Detail & Related papers (2022-05-25T12:41:59Z) - Explain to Not Forget: Defending Against Catastrophic Forgetting with
XAI [10.374979214803805]
Catastrophic forgetting describes the phenomenon when a neural network completely forgets previous knowledge when given new information.
We propose a novel training algorithm called training by explaining in which we leverage Layer-wise Relevance propagation in order to retain the information a neural network has already learned in previous tasks when training on new data.
Our method not only successfully retains the knowledge of old tasks within the neural networks but does so more resource-efficiently than other state-of-the-art solutions.
arXiv Detail & Related papers (2022-05-04T08:00:49Z) - The ideal data compression and automatic discovery of hidden law using
neural network [0.0]
We consider how the human brain recognizes events and memorizes them.
We reproduce the system of the human brain on a machine learning model with a new autoencoder neural network.
arXiv Detail & Related papers (2022-03-31T10:55:24Z) - Neurosymbolic hybrid approach to driver collision warning [64.02492460600905]
There are two main algorithmic approaches to autonomous driving systems.
Deep learning alone has achieved state-of-the-art results in many areas.
But sometimes it can be very difficult to debug if the deep learning model doesn't work.
arXiv Detail & Related papers (2022-03-28T20:29:50Z) - Ten Quick Tips for Deep Learning in Biology [116.78436313026478]
Machine learning is concerned with the development and applications of algorithms that can recognize patterns in data and use them for predictive modeling.
Deep learning has become its own subfield of machine learning.
In the context of biological research, deep learning has been increasingly used to derive novel insights from high-dimensional biological data.
arXiv Detail & Related papers (2021-05-29T21:02:44Z) - Towards a Universal Continuous Knowledge Base [49.95342223987143]
We propose a method for building a continuous knowledge base that can store knowledge imported from multiple neural networks.
Experiments on text classification show promising results.
We import the knowledge from multiple models to the knowledge base, from which the fused knowledge is exported back to a single model.
arXiv Detail & Related papers (2020-12-25T12:27:44Z) - Boosting Deep Neural Networks with Geometrical Prior Knowledge: A Survey [77.99182201815763]
Deep Neural Networks (DNNs) achieve state-of-the-art results in many different problem settings.
DNNs are often treated as black box systems, which complicates their evaluation and validation.
One promising field, inspired by the success of convolutional neural networks (CNNs) in computer vision tasks, is to incorporate knowledge about symmetric geometrical transformations.
arXiv Detail & Related papers (2020-06-30T14:56:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.