Evaluating the Progress of Deep Learning for Visual Relational Concepts
- URL: http://arxiv.org/abs/2001.10857v3
- Date: Mon, 13 Sep 2021 15:19:39 GMT
- Title: Evaluating the Progress of Deep Learning for Visual Relational Concepts
- Authors: Sebastian Stabinger, Peer David, Justus Piater, and Antonio
Rodr\'iguez-S\'anchez
- Abstract summary: We will show that difficult tasks are linked to relational concepts from cognitive psychology.
We will review research that is linked to relational concept learning, even if it was not originally presented from this angle.
We will recommend steps to make future datasets more relevant for testing systems on relational reasoning.
- Score: 0.6999740786886536
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional Neural Networks (CNNs) have become the state of the art method
for image classification in the last ten years. Despite the fact that they
achieve superhuman classification accuracy on many popular datasets, they often
perform much worse on more abstract image classification tasks. We will show
that these difficult tasks are linked to relational concepts from cognitive
psychology and that despite progress over the last few years, such relational
reasoning tasks still remain difficult for current neural network
architectures.
We will review deep learning research that is linked to relational concept
learning, even if it was not originally presented from this angle. Reviewing
the current literature, we will argue that some form of attention will be an
important component of future systems to solve relational tasks.
In addition, we will point out the shortcomings of currently used datasets,
and we will recommend steps to make future datasets more relevant for testing
systems on relational reasoning.
Related papers
- Development of Skip Connection in Deep Neural Networks for Computer Vision and Medical Image Analysis: A Survey [6.2299272077865675]
skip connection has played an essential role in the architecture of deep neural networks.
This survey provides a comprehensive summary and outlook on the development of skip connections in deep neural networks.
We summarize seminal papers, source code, models, and datasets that utilize skip connections in computer vision.
arXiv Detail & Related papers (2024-05-02T20:43:58Z) - Look-Ahead Selective Plasticity for Continual Learning of Visual Tasks [9.82510084910641]
We propose a new mechanism that takes place during task boundaries, i.e., when one task finishes and another starts.
We evaluate the proposed methods on benchmark computer vision datasets including CIFAR10 and TinyImagenet.
arXiv Detail & Related papers (2023-11-02T22:00:23Z) - Feature Forgetting in Continual Representation Learning [48.89340526235304]
representations do not suffer from "catastrophic forgetting" even in plain continual learning, but little further fact is known about its characteristics.
We devise a protocol for evaluating representation in continual learning, and then use it to present an overview of the basic trends of continual representation learning.
To study the feature forgetting problem, we create a synthetic dataset to identify and visualize the prevalence of feature forgetting in neural networks.
arXiv Detail & Related papers (2022-05-26T13:38:56Z) - Deep transfer learning for image classification: a survey [4.590533239391236]
Best practice for image classification is when large deep models can be trained on abundant labelled data.
In these scenarios transfer learning can help improve performance.
We present a new taxonomy of the applications of transfer learning for image classification.
arXiv Detail & Related papers (2022-05-20T00:03:39Z) - On Generalizing Beyond Domains in Cross-Domain Continual Learning [91.56748415975683]
Deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.
Our proposed approach learns new tasks under domain shift with accuracy boosts up to 10% on challenging datasets such as DomainNet and OfficeHome.
arXiv Detail & Related papers (2022-03-08T09:57:48Z) - Neural Architecture Search for Dense Prediction Tasks in Computer Vision [74.9839082859151]
Deep learning has led to a rising demand for neural network architecture engineering.
neural architecture search (NAS) aims at automatically designing neural network architectures in a data-driven manner rather than manually.
NAS has become applicable to a much wider range of problems in computer vision.
arXiv Detail & Related papers (2022-02-15T08:06:50Z) - Deep Long-Tailed Learning: A Survey [163.16874896812885]
Deep long-tailed learning aims to train well-performing deep models from a large number of images that follow a long-tailed class distribution.
Long-tailed class imbalance is a common problem in practical visual recognition tasks.
This paper provides a comprehensive survey on recent advances in deep long-tailed learning.
arXiv Detail & Related papers (2021-10-09T15:25:22Z) - What Is Considered Complete for Visual Recognition? [110.43159801737222]
We advocate for a new type of pre-training task named learning-by-compression.
The computational models are optimized to represent the visual data using compact features.
Semantic annotations, when available, play the role of weak supervision.
arXiv Detail & Related papers (2021-05-28T16:59:14Z) - Gradient Projection Memory for Continual Learning [5.43185002439223]
The ability to learn continually without forgetting the past tasks is a desired attribute for artificial learning systems.
We propose a novel approach where a neural network learns new tasks by taking gradient steps in the orthogonal direction to the gradient subspaces deemed important for the past tasks.
arXiv Detail & Related papers (2021-03-17T16:31:29Z) - Learning Contact Dynamics using Physically Structured Neural Networks [81.73947303886753]
We use connections between deep neural networks and differential equations to design a family of deep network architectures for representing contact dynamics between objects.
We show that these networks can learn discontinuous contact events in a data-efficient manner from noisy observations.
Our results indicate that an idealised form of touch feedback is a key component of making this learning problem tractable.
arXiv Detail & Related papers (2021-02-22T17:33:51Z) - Saliency Prediction with External Knowledge [27.75589849982756]
We develop a new Graph Semantic Saliency Network (GraSSNet) that constructs a graph that encodes semantic relationships learned from external knowledge.
A Spatial Graph Attention Network is then developed to update saliency features based on the learned graph.
Experiments show that the proposed model learns to predict saliency from the external knowledge and outperforms the state-of-the-art on four saliency benchmarks.
arXiv Detail & Related papers (2020-07-27T20:12:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.