Knowledge accumulating: The general pattern of learning
- URL: http://arxiv.org/abs/2108.03988v1
- Date: Mon, 9 Aug 2021 12:41:28 GMT
- Title: Knowledge accumulating: The general pattern of learning
- Authors: Zhuoran Xu and Hao Liu
- Abstract summary: In solving real world tasks, we still need to adjust algorithms to fit task unique features.
A single algorithm, no matter how we improve it, can only solve dense feedback tasks or specific sparse feedback tasks.
This paper first analyses how sparse feedback affects algorithm perfomance, and then proposes a pattern that explains how to accumulate knowledge to solve sparse feedback problems.
- Score: 5.174379158867218
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Artificial Intelligence has been developed for decades with the achievement
of great progress. Recently, deep learning shows its ability to solve many real
world problems, e.g. image classification and detection, natural language
processing, playing GO. Theoretically speaking, an artificial neural network
can fit any function and reinforcement learning can learn from any delayed
reward. But in solving real world tasks, we still need to spend a lot of effort
to adjust algorithms to fit task unique features. This paper proposes that the
reason of this phenomenon is the sparse feedback feature of the nature, and a
single algorithm, no matter how we improve it, can only solve dense feedback
tasks or specific sparse feedback tasks. This paper first analyses how sparse
feedback affects algorithm perfomance, and then proposes a pattern that
explains how to accumulate knowledge to solve sparse feedback problems.
Related papers
- The Clock and the Pizza: Two Stories in Mechanistic Explanation of
Neural Networks [59.26515696183751]
We show that algorithm discovery in neural networks is sometimes more complex.
We show that even simple learning problems can admit a surprising diversity of solutions.
arXiv Detail & Related papers (2023-06-30T17:59:13Z) - A Generalist Neural Algorithmic Learner [18.425083543441776]
We build a single graph neural network processor capable of learning to execute a wide range of algorithms.
We show that it is possible to effectively learn algorithms in a multi-task manner, so long as we can learn to execute them well in a single-task regime.
arXiv Detail & Related papers (2022-09-22T16:41:33Z) - Learning Iterative Reasoning through Energy Minimization [77.33859525900334]
We present a new framework for iterative reasoning with neural networks.
We train a neural network to parameterize an energy landscape over all outputs.
We implement each step of the iterative reasoning as an energy minimization step to find a minimal energy solution.
arXiv Detail & Related papers (2022-06-30T17:44:20Z) - Refining neural network predictions using background knowledge [68.35246878394702]
We show we can use logical background knowledge in learning system to compensate for a lack of labeled training data.
We introduce differentiable refinement functions that find a corrected prediction close to the original prediction.
This algorithm finds optimal refinements on complex SAT formulas in significantly fewer iterations and frequently finds solutions where gradient descent can not.
arXiv Detail & Related papers (2022-06-10T10:17:59Z) - Feature Forgetting in Continual Representation Learning [48.89340526235304]
representations do not suffer from "catastrophic forgetting" even in plain continual learning, but little further fact is known about its characteristics.
We devise a protocol for evaluating representation in continual learning, and then use it to present an overview of the basic trends of continual representation learning.
To study the feature forgetting problem, we create a synthetic dataset to identify and visualize the prevalence of feature forgetting in neural networks.
arXiv Detail & Related papers (2022-05-26T13:38:56Z) - End-to-end Algorithm Synthesis with Recurrent Networks: Logical
Extrapolation Without Overthinking [52.05847268235338]
We show how machine learning systems can perform logical extrapolation without overthinking problems.
We propose a recall architecture that keeps an explicit copy of the problem instance in memory so that it cannot be forgotten.
We also employ a progressive training routine that prevents the model from learning behaviors that are specific to number and instead pushes it to learn behaviors that can be repeated indefinitely.
arXiv Detail & Related papers (2022-02-11T18:43:28Z) - Some thoughts on catastrophic forgetting and how to learn an algorithm [0.0]
We propose to use a neural network with a different architecture that can be trained to recover the correct algorithm for the addition of binary numbers.
The neural network not only does not suffer from catastrophic forgetting but it improves its predictive power on unseen pairs of numbers as training progresses.
arXiv Detail & Related papers (2021-08-09T11:12:43Z) - Systematic human learning and generalization from a brief tutorial with
explanatory feedback [3.7826494079172557]
We investigate human adults' ability to learn an abstract reasoning task based on Sudoku.
We find that participants who master the task do so within a small number of trials and generalize well to puzzles outside of the training range.
We also find that most of those who master the task can describe a valid solution strategy, and such participants perform better on transfer puzzles than those whose strategy descriptions are vague or incomplete.
arXiv Detail & Related papers (2021-07-10T00:14:41Z) - Thinking Deeply with Recurrence: Generalizing from Easy to Hard
Sequential Reasoning Problems [51.132938969015825]
We observe that recurrent networks have the uncanny ability to closely emulate the behavior of non-recurrent deep models.
We show that recurrent networks that are trained to solve simple mazes with few recurrent steps can indeed solve much more complex problems simply by performing additional recurrences during inference.
arXiv Detail & Related papers (2021-02-22T14:09:20Z) - Demystifying Deep Neural Networks Through Interpretation: A Survey [3.566184392528658]
Modern deep learning algorithms tend to optimize an objective metric, such as minimize a cross entropy loss on a training dataset, to be able to learn.
The problem is that the single metric is an incomplete description of the real world tasks.
There are works done to tackle the problem of interpretability to provide insights into neural networks behavior and thought process.
arXiv Detail & Related papers (2020-12-13T17:56:41Z) - Curriculum Learning with Hindsight Experience Replay for Sequential
Object Manipulation Tasks [1.370633147306388]
We present an algorithm that combines curriculum learning with Hindsight Experience Replay (HER) to learn sequential object manipulation tasks.
The algorithm exploits the recurrent structure inherent in many object manipulation tasks and implements the entire learning process in the original simulation without adjusting it to each source task.
arXiv Detail & Related papers (2020-08-21T08:59:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.