Ultrafast Image Categorization in Biology and Neural Models
- URL: http://arxiv.org/abs/2205.03635v4
- Date: Wed, 31 May 2023 05:30:51 GMT
- Title: Ultrafast Image Categorization in Biology and Neural Models
- Authors: Jean-Nicolas J\'er\'emie, Laurent U Perrinet
- Abstract summary: We re-trained the standard VGG 16 CNN on two independent tasks that are ecologically relevant to humans.
We show that re-training the network achieves a human-like level of performance, comparable to that reported in psychophysical tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans are able to categorize images very efficiently, in particular to
detect the presence of an animal very quickly. Recently, deep learning
algorithms based on convolutional neural networks (CNNs) have achieved higher
than human accuracy for a wide range of visual categorization tasks. However,
the tasks on which these artificial networks are typically trained and
evaluated tend to be highly specialized and do not generalize well, e.g.,
accuracy drops after image rotation. In this respect, biological visual systems
are more flexible and efficient than artificial systems for more general tasks,
such as recognizing an animal. To further the comparison between biological and
artificial neural networks, we re-trained the standard VGG 16 CNN on two
independent tasks that are ecologically relevant to humans: detecting the
presence of an animal or an artifact. We show that re-training the network
achieves a human-like level of performance, comparable to that reported in
psychophysical tasks. In addition, we show that the categorization is better
when the outputs of the models are combined. Indeed, animals (e.g., lions) tend
to be less present in photographs that contain artifacts (e.g., buildings).
Furthermore, these re-trained models were able to reproduce some unexpected
behavioral observations from human psychophysics, such as robustness to
rotation (e.g., an upside-down or tilted image) or to a grayscale
transformation. Finally, we quantified the number of CNN layers required to
achieve such performance and showed that good accuracy for ultrafast image
categorization can be achieved with only a few layers, challenging the belief
that image recognition requires deep sequential analysis of visual objects.
Related papers
- Evaluating alignment between humans and neural network representations in image-based learning tasks [5.657101730705275]
We tested how well the representations of $86$ pretrained neural network models mapped to human learning trajectories.
We found that while training dataset size was a core determinant of alignment with human choices, contrastive training with multi-modal data (text and imagery) was a common feature of currently publicly available models that predicted human generalisation.
In conclusion, pretrained neural networks can serve to extract representations for cognitive models, as they appear to capture some fundamental aspects of cognition that are transferable across tasks.
arXiv Detail & Related papers (2023-06-15T08:18:29Z) - Multi-Domain Norm-referenced Encoding Enables Data Efficient Transfer
Learning of Facial Expression Recognition [62.997667081978825]
We propose a biologically-inspired mechanism for transfer learning in facial expression recognition.
Our proposed architecture provides an explanation for how the human brain might innately recognize facial expressions on varying head shapes.
Our model achieves a classification accuracy of 92.15% on the FERG dataset with extreme data efficiency.
arXiv Detail & Related papers (2023-04-05T09:06:30Z) - Connecting metrics for shape-texture knowledge in computer vision [1.7785095623975342]
Deep neural networks remain brittle and susceptible to many changes in the image that do not cause humans to misclassify images.
Part of this different behavior may be explained by the type of features humans and deep neural networks use in vision tasks.
arXiv Detail & Related papers (2023-01-25T14:37:42Z) - Wild Animal Classifier Using CNN [0.0]
Convolution neural networks (CNNs) have multiple layers which have different weights for the purpose of prediction of a particular input.
Image segmentation is one such widely used image processing method which provides a clear demarcation of the areas of interest in the image.
arXiv Detail & Related papers (2022-10-03T13:14:08Z) - Guiding Visual Attention in Deep Convolutional Neural Networks Based on
Human Eye Movements [0.0]
Deep Convolutional Neural Networks (DCNNs) were originally inspired by principles of biological vision.
Recent advances in deep learning seem to decrease this similarity.
We investigate a purely data-driven approach to obtain useful models.
arXiv Detail & Related papers (2022-06-21T17:59:23Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Overcoming the Domain Gap in Neural Action Representations [60.47807856873544]
3D pose data can now be reliably extracted from multi-view video sequences without manual intervention.
We propose to use it to guide the encoding of neural action representations together with a set of neural and behavioral augmentations.
To reduce the domain gap, during training, we swap neural and behavioral data across animals that seem to be performing similar actions.
arXiv Detail & Related papers (2021-12-02T12:45:46Z) - Continuous Emotion Recognition with Spatiotemporal Convolutional Neural
Networks [82.54695985117783]
We investigate the suitability of state-of-the-art deep learning architectures for continuous emotion recognition using long video sequences captured in-the-wild.
We have developed and evaluated convolutional recurrent neural networks combining 2D-CNNs and long short term-memory units, and inflated 3D-CNN models, which are built by inflating the weights of a pre-trained 2D-CNN model during fine-tuning.
arXiv Detail & Related papers (2020-11-18T13:42:05Z) - Fooling the primate brain with minimal, targeted image manipulation [67.78919304747498]
We propose an array of methods for creating minimal, targeted image perturbations that lead to changes in both neuronal activity and perception as reflected in behavior.
Our work shares the same goal with adversarial attack, namely the manipulation of images with minimal, targeted noise that leads ANN models to misclassify the images.
arXiv Detail & Related papers (2020-11-11T08:30:54Z) - Deep Low-Shot Learning for Biological Image Classification and
Visualization from Limited Training Samples [52.549928980694695]
In situ hybridization (ISH) gene expression pattern images from the same developmental stage are compared.
labeling training data with precise stages is very time-consuming even for biologists.
We propose a deep two-step low-shot learning framework to accurately classify ISH images using limited training images.
arXiv Detail & Related papers (2020-10-20T06:06:06Z) - Seeing eye-to-eye? A comparison of object recognition performance in
humans and deep convolutional neural networks under image manipulation [0.0]
This study aims towards a behavioral comparison of visual core object recognition performance between humans and feedforward neural networks.
Analyses of accuracy revealed that humans not only outperform DCNNs on all conditions, but also display significantly greater robustness towards shape and most notably color alterations.
arXiv Detail & Related papers (2020-07-13T10:26:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.