Design of Explainability Module with Experts in the Loop for
Visualization and Dynamic Adjustment of Continual Learning
- URL: http://arxiv.org/abs/2202.06781v1
- Date: Mon, 14 Feb 2022 15:00:22 GMT
- Title: Design of Explainability Module with Experts in the Loop for
Visualization and Dynamic Adjustment of Continual Learning
- Authors: Yujiang He, Zhixin Huang and Bernhard Sick
- Abstract summary: Continual learning can enable neural networks to evolve by learning new tasks sequentially in task-changing scenarios.
New novelties from the data stream in applications could contain anomalies that are meaningless for continual learning.
We propose the conceptual design of an explainability module with experts in the loop based on techniques, such as dimension reduction, visualization, and evaluation strategies.
- Score: 5.039779583329608
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Continual learning can enable neural networks to evolve by learning new tasks
sequentially in task-changing scenarios. However, two general and related
challenges should be overcome in further research before we apply this
technique to real-world applications. Firstly, newly collected novelties from
the data stream in applications could contain anomalies that are meaningless
for continual learning. Instead of viewing them as a new task for updating, we
have to filter out such anomalies to reduce the disturbance of extremely
high-entropy data for the progression of convergence. Secondly, fewer efforts
have been put into research regarding the explainability of continual learning,
which leads to a lack of transparency and credibility of the updated neural
networks. Elaborated explanations about the process and result of continual
learning can help experts in judgment and making decisions. Therefore, we
propose the conceptual design of an explainability module with experts in the
loop based on techniques, such as dimension reduction, visualization, and
evaluation strategies. This work aims to overcome the mentioned challenges by
sufficiently explaining and visualizing the identified anomalies and the
updated neural network. With the help of this module, experts can be more
confident in decision-making regarding anomaly filtering, dynamic adjustment of
hyperparameters, data backup, etc.
Related papers
- Deep Learning Through A Telescoping Lens: A Simple Model Provides Empirical Insights On Grokking, Gradient Boosting & Beyond [61.18736646013446]
In pursuit of a deeper understanding of its surprising behaviors, we investigate the utility of a simple yet accurate model of a trained neural network.
Across three case studies, we illustrate how it can be applied to derive new empirical insights on a diverse range of prominent phenomena.
arXiv Detail & Related papers (2024-10-31T22:54:34Z) - Temporal-Difference Variational Continual Learning [89.32940051152782]
A crucial capability of Machine Learning models in real-world applications is the ability to continuously learn new tasks.
In Continual Learning settings, models often struggle to balance learning new tasks with retaining previous knowledge.
We propose new learning objectives that integrate the regularization effects of multiple previous posterior estimations.
arXiv Detail & Related papers (2024-10-10T10:58:41Z) - Function-space Parameterization of Neural Networks for Sequential Learning [22.095632118886225]
Sequential learning paradigms pose challenges for gradient-based deep learning due to difficulties incorporating new data and retaining prior knowledge.
We introduce a technique that converts neural networks from weight space to function space, through a dual parameterization.
Our experiments demonstrate that we can retain knowledge in continual learning and incorporate new data efficiently.
arXiv Detail & Related papers (2024-03-16T14:00:04Z) - When Meta-Learning Meets Online and Continual Learning: A Survey [39.53836535326121]
meta-learning is a data-driven approach to optimize the learning algorithm.
Continual learning and online learning, both of which involve incrementally updating a model with streaming data.
This paper organizes various problem settings using consistent terminology and formal descriptions.
arXiv Detail & Related papers (2023-11-09T09:49:50Z) - Advancing continual lifelong learning in neural information retrieval: definition, dataset, framework, and empirical evaluation [3.2340528215722553]
A systematic task formulation of continual neural information retrieval is presented.
A comprehensive continual neural information retrieval framework is proposed.
Empirical evaluations illustrate that the proposed framework can successfully prevent catastrophic forgetting in neural information retrieval.
arXiv Detail & Related papers (2023-08-16T14:01:25Z) - Synergistic information supports modality integration and flexible
learning in neural networks solving multiple tasks [107.8565143456161]
We investigate the information processing strategies adopted by simple artificial neural networks performing a variety of cognitive tasks.
Results show that synergy increases as neural networks learn multiple diverse tasks.
randomly turning off neurons during training through dropout increases network redundancy, corresponding to an increase in robustness.
arXiv Detail & Related papers (2022-10-06T15:36:27Z) - Learning Dynamics and Generalization in Reinforcement Learning [59.530058000689884]
We show theoretically that temporal difference learning encourages agents to fit non-smooth components of the value function early in training.
We show that neural networks trained using temporal difference algorithms on dense reward tasks exhibit weaker generalization between states than randomly networks and gradient networks trained with policy methods.
arXiv Detail & Related papers (2022-06-05T08:49:16Z) - Online Deep Learning from Doubly-Streaming Data [17.119725174036653]
This paper investigates a new online learning problem with doubly-streaming data, where the data streams are described by feature spaces that constantly evolve.
A plausible idea to overcome the challenges is to establish relationship between the pre-and-post evolving feature spaces.
We propose a novel OLD3S paradigm, where a shared latent subspace is discovered to summarize information from the old and new feature spaces.
arXiv Detail & Related papers (2022-04-25T17:06:39Z) - Adaptive Explainable Continual Learning Framework for Regression
Problems with Focus on Power Forecasts [0.0]
Two continual learning scenarios will be proposed to describe the potential challenges in this context.
Deep neural networks have to learn new tasks and overcome forgetting the knowledge obtained from the old tasks as the amount of data keeps increasing in applications.
Research topics are related but not limited to developing continual deep learning algorithms, strategies for non-stationarity detection in data streams, explainable and visualizable artificial intelligence, etc.
arXiv Detail & Related papers (2021-08-24T14:59:10Z) - Causal Navigation by Continuous-time Neural Networks [108.84958284162857]
We propose a theoretical and experimental framework for learning causal representations using continuous-time neural networks.
We evaluate our method in the context of visual-control learning of drones over a series of complex tasks.
arXiv Detail & Related papers (2021-06-15T17:45:32Z) - Binary Neural Networks: A Survey [126.67799882857656]
The binary neural network serves as a promising technique for deploying deep models on resource-limited devices.
The binarization inevitably causes severe information loss, and even worse, its discontinuity brings difficulty to the optimization of the deep network.
We present a survey of these algorithms, mainly categorized into the native solutions directly conducting binarization, and the optimized ones using techniques like minimizing the quantization error, improving the network loss function, and reducing the gradient error.
arXiv Detail & Related papers (2020-03-31T16:47:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.