Controlled-rearing studies of newborn chicks and deep neural networks
- URL: http://arxiv.org/abs/2112.06106v1
- Date: Sun, 12 Dec 2021 00:45:07 GMT
- Title: Controlled-rearing studies of newborn chicks and deep neural networks
- Authors: Donsuk Lee, Pranav Gujarathi, Justin N. Wood
- Abstract summary: Convolutional neural networks (CNNs) can achieve human-level performance on challenging object recognition tasks.
CNNs are thought to be "data hungry," requiring massive amounts of training data to develop accurate models for object recognition.
This critique challenges the promise of using CNNs as models of visual development.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional neural networks (CNNs) can now achieve human-level performance
on challenging object recognition tasks. CNNs are also the leading quantitative
models in terms of predicting neural and behavioral responses in visual
recognition tasks. However, there is a widely accepted critique of CNN models:
unlike newborn animals, which learn rapidly and efficiently, CNNs are thought
to be "data hungry," requiring massive amounts of training data to develop
accurate models for object recognition. This critique challenges the promise of
using CNNs as models of visual development. Here, we directly examined whether
CNNs are more data hungry than newborn animals by performing parallel
controlled-rearing experiments on newborn chicks and CNNs. We raised newborn
chicks in strictly controlled visual environments, then simulated the training
data available in that environment by constructing a virtual animal chamber in
a video game engine. We recorded the visual images acquired by an agent moving
through the virtual chamber and used those images to train CNNs. When CNNs
received similar visual training data as chicks, the CNNs successfully solved
the same challenging view-invariant object recognition tasks as the chicks.
Thus, the CNNs were not more data hungry than animals: both CNNs and chicks
successfully developed robust object models from training data of a single
object.
Related papers
- Using Human-like Mechanism to Weaken Effect of Pre-training Weight Bias
in Face-Recognition Convolutional Neural Network [6.0950431324191845]
We focus on 4 extensively studied CNNs (AlexNet, VGG11, VGG13, and VGG16) which has been analyzed as human-like models by neuroscientists.
We trained these CNNs to emotion valence classification task by transfer learning.
We then update the object-based AlexNet using self-attention mechanism based on neuroscience and behavioral data.
arXiv Detail & Related papers (2023-10-20T17:22:57Z) - Transferability of Convolutional Neural Networks in Stationary Learning
Tasks [96.00428692404354]
We introduce a novel framework for efficient training of convolutional neural networks (CNNs) for large-scale spatial problems.
We show that a CNN trained on small windows of such signals achieves a nearly performance on much larger windows without retraining.
Our results show that the CNN is able to tackle problems with many hundreds of agents after being trained with fewer than ten.
arXiv Detail & Related papers (2023-07-21T13:51:45Z) - A novel feature-scrambling approach reveals the capacity of
convolutional neural networks to learn spatial relations [0.0]
Convolutional neural networks (CNNs) are one of the most successful computer vision systems to solve object recognition.
Yet it remains poorly understood how CNNs actually make their decisions, what the nature of their internal representations is, and how their recognition strategies differ from humans.
arXiv Detail & Related papers (2022-12-12T16:40:29Z) - InternImage: Exploring Large-Scale Vision Foundation Models with
Deformable Convolutions [95.94629864981091]
This work presents a new large-scale CNN-based foundation model, termed InternImage, which can obtain the gain from increasing parameters and training data like ViTs.
The proposed InternImage reduces the strict inductive bias of traditional CNNs and makes it possible to learn stronger and more robust patterns with large-scale parameters from massive data like ViTs.
arXiv Detail & Related papers (2022-11-10T18:59:04Z) - Improving the Accuracy and Robustness of CNNs Using a Deep CCA Neural
Data Regularizer [2.026424957803652]
As convolutional neural networks (CNNs) become more accurate at object recognition, their representations become more similar to the primate visual system.
Previous attempts to address this question showed very modest gains in accuracy, owing in part to limitations of the regularization method.
We develop a new neural data regularizer for CNNs that uses Deep Correlation Analysis (DCCA) to optimize the resemblance of the CNN's image representations to that of the monkey visual cortex.
arXiv Detail & Related papers (2022-09-06T15:40:39Z) - Prune and distill: similar reformatting of image information along rat
visual cortex and deep neural networks [61.60177890353585]
Deep convolutional neural networks (CNNs) have been shown to provide excellent models for its functional analogue in the brain, the ventral stream in visual cortex.
Here we consider some prominent statistical patterns that are known to exist in the internal representations of either CNNs or the visual cortex.
We show that CNNs and visual cortex share a similarly tight relationship between dimensionality expansion/reduction of object representations and reformatting of image information.
arXiv Detail & Related papers (2022-05-27T08:06:40Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - A CNN-based Feature Space for Semi-supervised Incremental Learning in
Assisted Living Applications [2.1485350418225244]
We propose using the feature space that results from the training dataset to automatically label problematic images.
The resulting semi-supervised incremental learning process allows improving the classification accuracy of new instances by 40%.
arXiv Detail & Related papers (2020-11-11T12:31:48Z) - An Information-theoretic Visual Analysis Framework for Convolutional
Neural Networks [11.15523311079383]
We introduce a data model to organize the data that can be extracted from CNN models.
We then propose two ways to calculate entropy under different circumstances.
We develop a visual analysis system, CNNSlicer, to interactively explore the amount of information changes inside the model.
arXiv Detail & Related papers (2020-05-02T21:36:50Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.