Extreme Image Transformations Affect Humans and Machines Differently
- URL: http://arxiv.org/abs/2212.13967v2
- Date: Tue, 11 Apr 2023 18:58:50 GMT
- Title: Extreme Image Transformations Affect Humans and Machines Differently
- Authors: Girik Malik and Dakarai Crowder and Ennio Mingolla
- Abstract summary: Some recent artificial neural networks (ANNs) claim to model aspects of primate neural and human performance data.
We introduce a set of novel image transforms inspired by neurophysiological findings and evaluate humans and ANNs on an object recognition task.
We show that machines perform better than humans for certain transforms and struggle to perform at par with humans on others that are easy for humans.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Some recent artificial neural networks (ANNs) claim to model aspects of
primate neural and human performance data. Their success in object recognition
is, however, dependent on exploiting low-level features for solving visual
tasks in a way that humans do not. As a result, out-of-distribution or
adversarial input is often challenging for ANNs. Humans instead learn abstract
patterns and are mostly unaffected by many extreme image distortions. We
introduce a set of novel image transforms inspired by neurophysiological
findings and evaluate humans and ANNs on an object recognition task. We show
that machines perform better than humans for certain transforms and struggle to
perform at par with humans on others that are easy for humans. We quantify the
differences in accuracy for humans and machines and find a ranking of
difficulty for our transforms for human data. We also suggest how certain
characteristics of human visual processing can be adapted to improve the
performance of ANNs for our difficult-for-machines transforms.
Related papers
- HINT: Learning Complete Human Neural Representations from Limited Viewpoints [69.76947323932107]
We propose a NeRF-based algorithm able to learn a detailed and complete human model from limited viewing angles.
As a result, our method can reconstruct complete humans even from a few viewing angles, increasing performance by more than 15% PSNR.
arXiv Detail & Related papers (2024-05-30T05:43:09Z) - UniHuman: A Unified Model for Editing Human Images in the Wild [49.896715833075106]
We propose UniHuman, a unified model that addresses multiple facets of human image editing in real-world settings.
To enhance the model's generation quality and generalization capacity, we leverage guidance from human visual encoders.
In user studies, UniHuman is preferred by the users in an average of 77% of cases.
arXiv Detail & Related papers (2023-12-22T05:00:30Z) - Neural Novel Actor: Learning a Generalized Animatable Neural
Representation for Human Actors [98.24047528960406]
We propose a new method for learning a generalized animatable neural representation from a sparse set of multi-view imagery of multiple persons.
The learned representation can be used to synthesize novel view images of an arbitrary person from a sparse set of cameras, and further animate them with the user's pose control.
arXiv Detail & Related papers (2022-08-25T07:36:46Z) - Guiding Visual Attention in Deep Convolutional Neural Networks Based on
Human Eye Movements [0.0]
Deep Convolutional Neural Networks (DCNNs) were originally inspired by principles of biological vision.
Recent advances in deep learning seem to decrease this similarity.
We investigate a purely data-driven approach to obtain useful models.
arXiv Detail & Related papers (2022-06-21T17:59:23Z) - Robustness of Humans and Machines on Object Recognition with Extreme
Image Transformations [0.0]
We introduce a novel set of image transforms and evaluate humans and networks on an object recognition task.
We found performance for a few common networks quickly decreases while humans are able to recognize objects with a high accuracy.
arXiv Detail & Related papers (2022-05-09T17:15:54Z) - Ultrafast Image Categorization in Biology and Neural Models [0.0]
We re-trained the standard VGG 16 CNN on two independent tasks that are ecologically relevant to humans.
We show that re-training the network achieves a human-like level of performance, comparable to that reported in psychophysical tasks.
arXiv Detail & Related papers (2022-05-07T11:19:40Z) - Overcoming the Domain Gap in Neural Action Representations [60.47807856873544]
3D pose data can now be reliably extracted from multi-view video sequences without manual intervention.
We propose to use it to guide the encoding of neural action representations together with a set of neural and behavioral augmentations.
To reduce the domain gap, during training, we swap neural and behavioral data across animals that seem to be performing similar actions.
arXiv Detail & Related papers (2021-12-02T12:45:46Z) - Where is my hand? Deep hand segmentation for visual self-recognition in
humanoid robots [129.46920552019247]
We propose the use of a Convolution Neural Network (CNN) to segment the robot hand from an image in an egocentric view.
We fine-tuned the Mask-RCNN network for the specific task of segmenting the hand of the humanoid robot Vizzy.
arXiv Detail & Related papers (2021-02-09T10:34:32Z) - Fooling the primate brain with minimal, targeted image manipulation [67.78919304747498]
We propose an array of methods for creating minimal, targeted image perturbations that lead to changes in both neuronal activity and perception as reflected in behavior.
Our work shares the same goal with adversarial attack, namely the manipulation of images with minimal, targeted noise that leads ANN models to misclassify the images.
arXiv Detail & Related papers (2020-11-11T08:30:54Z) - Deformable Convolutional LSTM for Human Body Emotion Recognition [0.0]
We do experiments on the GEM dataset and achieved state-of-the-art accuracy of 98.8% on the task of whole human body emotion recognition.
We incorporate the deformable behavior into the core of convolutional long short-term memory (ConvLSTM) to improve robustness to these deformations in the image.
arXiv Detail & Related papers (2020-10-27T21:01:09Z) - Salient Facial Features from Humans and Deep Neural Networks [2.5211876507510724]
We explore the features that are used by humans and by convolutional neural networks (ConvNets) to classify faces.
We use Guided Backpropagation (GB) to visualize the facial features that influence the output of a ConvNet the most when identifying specific individuals.
arXiv Detail & Related papers (2020-03-08T22:41:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.