Self-Supervised Learning of a Biologically-Inspired Visual Texture Model
- URL: http://arxiv.org/abs/2006.16976v1
- Date: Tue, 30 Jun 2020 17:12:09 GMT
- Title: Self-Supervised Learning of a Biologically-Inspired Visual Texture Model
- Authors: Nikhil Parthasarathy and Eero P. Simoncelli
- Abstract summary: We develop a model for representing visual texture in a low-dimensional feature space.
Inspired by the architecture of primate visual cortex, the model uses a first stage of oriented linear filters.
We show that the learned model exhibits stronger representational similarity to texture responses of neural populations recorded in primate V2 than pre-trained deep CNNs.
- Score: 6.931125029302013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop a model for representing visual texture in a low-dimensional
feature space, along with a novel self-supervised learning objective that is
used to train it on an unlabeled database of texture images. Inspired by the
architecture of primate visual cortex, the model uses a first stage of oriented
linear filters (corresponding to cortical area V1), consisting of both
rectified units (simple cells) and pooled phase-invariant units (complex
cells). These responses are processed by a second stage (analogous to cortical
area V2) consisting of convolutional filters followed by half-wave
rectification and pooling to generate V2 'complex cell' responses. The second
stage filters are trained on a set of unlabeled homogeneous texture images,
using a novel contrastive objective that maximizes the distance between the
distribution of V2 responses to individual images and the distribution of
responses across all images. When evaluated on texture classification, the
trained model achieves substantially greater data-efficiency than a variety of
deep hierarchical model architectures. Moreover, we show that the learned model
exhibits stronger representational similarity to texture responses of neural
populations recorded in primate V2 than pre-trained deep CNNs.
Related papers
- pAE: An Efficient Autoencoder Architecture for Modeling the Lateral Geniculate Nucleus by Integrating Feedforward and Feedback Streams in Human Visual System [0.716879432974126]
We introduce a deep convolutional model that closely approximates human visual information processing.
We aim to approximate the function for the lateral geniculate nucleus (LGN) area using a trained shallow convolutional model.
The pAE model achieves the final 99.26% prediction performance and demonstrates a notable improvement of around 28% over human results in the temporal mode.
arXiv Detail & Related papers (2024-09-20T16:33:01Z) - Layerwise complexity-matched learning yields an improved model of cortical area V2 [12.861402235256207]
Deep neural networks trained end-to-end for object recognition approach human capabilities.
We develop a self-supervised training methodology that operates independently on successive layers.
We show that our model is better aligned with selectivity properties and neural activity in primate area V2.
arXiv Detail & Related papers (2023-12-18T18:37:02Z) - Improving Human-Object Interaction Detection via Virtual Image Learning [68.56682347374422]
Human-Object Interaction (HOI) detection aims to understand the interactions between humans and objects.
In this paper, we propose to alleviate the impact of such an unbalanced distribution via Virtual Image Leaning (VIL)
A novel label-to-image approach, Multiple Steps Image Creation (MUSIC), is proposed to create a high-quality dataset that has a consistent distribution with real images.
arXiv Detail & Related papers (2023-08-04T10:28:48Z) - Parents and Children: Distinguishing Multimodal DeepFakes from Natural Images [60.34381768479834]
Recent advancements in diffusion models have enabled the generation of realistic deepfakes from textual prompts in natural language.
We pioneer a systematic study on deepfake detection generated by state-of-the-art diffusion models.
arXiv Detail & Related papers (2023-04-02T10:25:09Z) - DeepDC: Deep Distance Correlation as a Perceptual Image Quality
Evaluator [53.57431705309919]
ImageNet pre-trained deep neural networks (DNNs) show notable transferability for building effective image quality assessment (IQA) models.
We develop a novel full-reference IQA (FR-IQA) model based exclusively on pre-trained DNN features.
We conduct comprehensive experiments to demonstrate the superiority of the proposed quality model on five standard IQA datasets.
arXiv Detail & Related papers (2022-11-09T14:57:27Z) - Top-down inference in an early visual cortex inspired hierarchical
Variational Autoencoder [0.0]
We exploit advances in Variational Autoencoders to investigate the early visual cortex with sparse coding hierarchical VAEs trained on natural images.
We show that representations similar to the one found in the primary and secondary visual cortices naturally emerge under mild inductive biases.
We show that a neuroscience-inspired choice of the recognition model is critical for two signatures of computations with generative models.
arXiv Detail & Related papers (2022-06-01T12:21:58Z) - Deep Learning Based Automated COVID-19 Classification from Computed
Tomography Images [0.0]
The paper presents a Convolutional Neural Networks (CNN) model for image classification, aiming at increasing predictive performance for COVID-19 diagnosis.
This work proposes a less complex solution based on simply classifying 2D CT-Scan slices of images using their pixels via a 2D CNN model.
Despite the simplicity in architecture, the proposed model showed improved quantitative results exceeding state-of-the-art on the same dataset of images.
arXiv Detail & Related papers (2021-11-22T13:35:10Z) - Emergence of Lie symmetries in functional architectures learned by CNNs [63.69764116066748]
We study the spontaneous development of symmetries in the early layers of a Convolutional Neural Network (CNN) during learning on natural images.
Our architecture is built in such a way to mimic the early stages of biological visual systems.
arXiv Detail & Related papers (2021-04-17T13:23:26Z) - Generating Diverse Structure for Image Inpainting With Hierarchical
VQ-VAE [74.29384873537587]
We propose a two-stage model for diverse inpainting, where the first stage generates multiple coarse results each of which has a different structure, and the second stage refines each coarse result separately by augmenting texture.
Experimental results on CelebA-HQ, Places2, and ImageNet datasets show that our method not only enhances the diversity of the inpainting solutions but also improves the visual quality of the generated multiple images.
arXiv Detail & Related papers (2021-03-18T05:10:49Z) - Emergent Properties of Foveated Perceptual Systems [3.3504365823045044]
This work is inspired by the foveated human visual system, which has higher acuity at the center of gaze and texture-like encoding in the periphery.
We introduce models consisting of a first-stage textitfixed image transform followed by a second-stage textitlearnable convolutional neural network.
We find that foveation with peripheral texture-based computations yields an efficient, distinct, and robust representational format of scene information.
arXiv Detail & Related papers (2020-06-14T19:34:44Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.