Superpixel-based Domain-Knowledge Infusion in Computer Vision
- URL: http://arxiv.org/abs/2105.09448v1
- Date: Thu, 20 May 2021 01:25:42 GMT
- Title: Superpixel-based Domain-Knowledge Infusion in Computer Vision
- Authors: Gunjan Chhablani, Abheesht Sharma, Harshit Pandey, Tirtharaj Dash
- Abstract summary: Superpixels are higher-order perceptual groups of pixels in an image, often carrying much more information than raw pixels.
There is an inherent relational structure to the relationship among different superpixels of an image.
This relational information can convey some form of domain information about the image, e.g. relationship between superpixels representing two eyes in a cat image.
- Score: 0.7349727826230862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Superpixels are higher-order perceptual groups of pixels in an image, often
carrying much more information than raw pixels. There is an inherent relational
structure to the relationship among different superpixels of an image. This
relational information can convey some form of domain information about the
image, e.g. relationship between superpixels representing two eyes in a cat
image. Our interest in this paper is to construct computer vision models,
specifically those based on Deep Neural Networks (DNNs) to incorporate these
superpixels information. We propose a methodology to construct a hybrid model
that leverages (a) Convolutional Neural Network (CNN) to deal with spatial
information in an image, and (b) Graph Neural Network (GNN) to deal with
relational superpixel information in the image. The proposed deep model is
learned using a generic hybrid loss function that we call a `hybrid' loss. We
evaluate the predictive performance of our proposed hybrid vision model on four
popular image classification datasets: MNIST, FMNIST, CIFAR-10 and CIFAR-100.
Moreover, we evaluate our method on three real-world classification tasks:
COVID-19 X-Ray Detection, LFW Face Recognition, and SOCOFing Fingerprint
Identification. The results demonstrate that the relational superpixel
information provided via a GNN could improve the performance of standard
CNN-based vision systems.
Related papers
- T-former: An Efficient Transformer for Image Inpainting [50.43302925662507]
A class of attention-based network architectures, called transformer, has shown significant performance on natural language processing fields.
In this paper, we design a novel attention linearly related to the resolution according to Taylor expansion, and based on this attention, a network called $T$-former is designed for image inpainting.
Experiments on several benchmark datasets demonstrate that our proposed method achieves state-of-the-art accuracy while maintaining a relatively low number of parameters and computational complexity.
arXiv Detail & Related papers (2023-05-12T04:10:42Z) - Single-Image Super-Resolution Reconstruction based on the Differences of
Neighboring Pixels [3.257500143434429]
The deep learning technique was used to increase the performance of single image super-resolution (SISR)
In this paper, we propose the differences of neighboring pixels to regularize the CNN by constructing a graph from the estimated image and the ground-truth image.
The proposed method outperforms the state-of-the-art methods in terms of quantitative and qualitative evaluation of the benchmark datasets.
arXiv Detail & Related papers (2022-12-28T07:30:07Z) - Traditional Classification Neural Networks are Good Generators: They are
Competitive with DDPMs and GANs [104.72108627191041]
We show that conventional neural network classifiers can generate high-quality images comparable to state-of-the-art generative models.
We propose a mask-based reconstruction module to make semantic gradients-aware to synthesize plausible images.
We show that our method is also applicable to text-to-image generation by regarding image-text foundation models.
arXiv Detail & Related papers (2022-11-27T11:25:35Z) - Unsupervised Image Semantic Segmentation through Superpixels and Graph
Neural Networks [6.123324869194195]
Unsupervised image segmentation is an important task in many real-world scenarios where labelled data is of scarce availability.
We propose a novel approach that harnesses recent advances in unsupervised learning using a combination of Mutual Information Maximization (MIM), Neural Superpixel and Graph Neural Networks (GNNs) in an end-to-end manner.
arXiv Detail & Related papers (2022-10-21T08:35:18Z) - Rethinking Unsupervised Neural Superpixel Segmentation [6.123324869194195]
unsupervised learning for superpixel segmentation via CNNs has been studied.
We propose three key elements to improve the efficacy of such networks.
By experimenting with the BSDS500 dataset, we find evidence to the significance of our proposal.
arXiv Detail & Related papers (2022-06-21T09:30:26Z) - Image Super-resolution with An Enhanced Group Convolutional Neural
Network [102.2483249598621]
CNNs with strong learning abilities are widely chosen to resolve super-resolution problem.
We present an enhanced super-resolution group CNN (ESRGCNN) with a shallow architecture.
Experiments report that our ESRGCNN surpasses the state-of-the-arts in terms of SISR performance, complexity, execution speed, image quality evaluation and visual effect in SISR.
arXiv Detail & Related papers (2022-05-29T00:34:25Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Implicit Integration of Superpixel Segmentation into Fully Convolutional
Networks [11.696069523681178]
We propose a way to implicitly integrate a superpixel scheme into CNNs.
Our proposed method hierarchically groups pixels at downsampling layers and generates superpixels.
We evaluate our method on several tasks such as semantic segmentation, superpixel segmentation, and monocular depth estimation.
arXiv Detail & Related papers (2021-03-05T02:20:26Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - AINet: Association Implantation for Superpixel Segmentation [82.21559299694555]
We propose a novel textbfAssociation textbfImplantation (AI) module to enable the network to explicitly capture the relations between the pixel and its surrounding grids.
Our method could not only achieve state-of-the-art performance but maintain satisfactory inference efficiency.
arXiv Detail & Related papers (2021-01-26T10:40:13Z) - Superpixel Image Classification with Graph Attention Networks [4.714325419968082]
This paper presents a methodology for image classification using Graph Neural Network (GNN) models.
We transform the input images into region adjacency graphs (RAGs), in which regions are superpixels and edges connect neighboring superpixels.
Experiments suggest that Graph Attention Networks (GATs), which combine graph convolutions with self-attention mechanisms, outperforms other GNN models.
arXiv Detail & Related papers (2020-02-13T14:52:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.