Learning Spatial Relationships between Samples of Patent Image Shapes
- URL: http://arxiv.org/abs/2004.05713v3
- Date: Mon, 27 Apr 2020 14:26:30 GMT
- Title: Learning Spatial Relationships between Samples of Patent Image Shapes
- Authors: Juan Castorena, Manish Bhattarai, Diane Oyen
- Abstract summary: We propose a method suitable to binary images which bridges some of the successes of deep learning (DL)
The method consists on extracting the shape of interest from the binary image and applying a non-Euclidean geometric neural-net architecture to learn the local and global spatial relationships of the shape.
- Score: 14.37369942979269
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Binary image based classification and retrieval of documents of an
intellectual nature is a very challenging problem. Variations in the binary
image generation mechanisms which are subject to the document artisan designer
including drawing style, view-point, inclusion of multiple image components are
plausible causes for increasing the complexity of the problem. In this work, we
propose a method suitable to binary images which bridges some of the successes
of deep learning (DL) to alleviate the problems introduced by the
aforementioned variations. The method consists on extracting the shape of
interest from the binary image and applying a non-Euclidean geometric
neural-net architecture to learn the local and global spatial relationships of
the shape. Empirical results show that our method is in some sense invariant to
the image generation mechanism variations and achieves results outperforming
existing methods in a patent image dataset benchmark.
Related papers
- Bayesian Unsupervised Disentanglement of Anatomy and Geometry for Deep Groupwise Image Registration [50.62725807357586]
This article presents a general Bayesian learning framework for multi-modal groupwise image registration.
We propose a novel hierarchical variational auto-encoding architecture to realise the inference procedure of the latent variables.
Experiments were conducted to validate the proposed framework, including four different datasets from cardiac, brain, and abdominal medical images.
arXiv Detail & Related papers (2024-01-04T08:46:39Z) - Detecting Generated Images by Real Images Only [64.12501227493765]
Existing generated image detection methods detect visual artifacts in generated images or learn discriminative features from both real and generated images by massive training.
This paper approaches the generated image detection problem from a new perspective: Start from real images.
By finding the commonality of real images and mapping them to a dense subspace in feature space, the goal is that generated images, regardless of their generative model, are then projected outside the subspace.
arXiv Detail & Related papers (2023-11-02T03:09:37Z) - Object-Centric Relational Representations for Image Generation [18.069747511100132]
This paper explores a novel method to condition image generation, based on object-centric relational representations.
We show that such architectural biases entail properties that facilitate the manipulation and conditioning of the generative process.
We also propose a novel benchmark for image generation consisting of a synthetic dataset of images paired with their relational representation.
arXiv Detail & Related papers (2023-03-26T11:17:17Z) - Share With Thy Neighbors: Single-View Reconstruction by Cross-Instance
Consistency [59.427074701985795]
Single-view reconstruction typically rely on viewpoint annotations, silhouettes, the absence of background, multiple views of the same instance, a template shape, or symmetry.
We avoid all of these supervisions and hypotheses by leveraging explicitly the consistency between images of different object instances.
Our main contributions are two approaches to leverage cross-instance consistency: (i) progressive conditioning, a training strategy to gradually specialize the model from category to instances in a curriculum learning fashion; (ii) swap reconstruction, a loss enforcing consistency between instances having similar shape or texture.
arXiv Detail & Related papers (2022-04-21T17:47:35Z) - A Survey on Intrinsic Images: Delving Deep Into Lambert and Beyond [8.313161485540338]
Intrinsic imaging or intrinsic image decomposition has traditionally been described as the problem of decomposing an image into two layers.
Deep learning techniques have been broadly applied in recent years to increase the accuracy of those separations.
We show that there is increasing awareness on the potential of more sophisticated physically-principled components of the image formation process.
arXiv Detail & Related papers (2021-12-07T17:26:35Z) - Learning Contrastive Representation for Semantic Correspondence [150.29135856909477]
We propose a multi-level contrastive learning approach for semantic matching.
We show that image-level contrastive learning is a key component to encourage the convolutional features to find correspondence between similar objects.
arXiv Detail & Related papers (2021-09-22T18:34:14Z) - An Optical physics inspired CNN approach for intrinsic image
decomposition [0.0]
Intrinsic Image Decomposition is an open problem of generating the constituents of an image.
We propose a neural network architecture capable of this decomposition using physics-based parameters derived from the image.
arXiv Detail & Related papers (2021-05-21T00:54:01Z) - Learning Deformable Image Registration from Optimization: Perspective,
Modules, Bilevel Training and Beyond [62.730497582218284]
We develop a new deep learning based framework to optimize a diffeomorphic model via multi-scale propagation.
We conduct two groups of image registration experiments on 3D volume datasets including image-to-atlas registration on brain MRI data and image-to-image registration on liver CT data.
arXiv Detail & Related papers (2020-04-30T03:23:45Z) - Diagram Image Retrieval using Sketch-Based Deep Learning and Transfer
Learning [28.31395148295604]
This paper presents a deep learning based method for image-based search for binary patent images.
We begin by using deep learning to generate sketches from natural images for image retrieval.
We then use our small set of manually labeled patent diagram images via transfer learning to adapt the image search from sketches of natural images to diagrams.
arXiv Detail & Related papers (2020-04-22T18:27:46Z) - Learning Transformation-Aware Embeddings for Image Forensics [15.484408315588569]
Image Provenance Analysis aims at discovering relationships among different manipulated image versions that share content.
One of the main sub-problems for provenance analysis that has not yet been addressed directly is the edit ordering of images that share full content or are near-duplicates.
This paper introduces a novel deep learning-based approach to provide a plausible ordering to images that have been generated from a single image through transformations.
arXiv Detail & Related papers (2020-01-13T22:01:24Z) - Fine-grained Image-to-Image Transformation towards Visual Recognition [102.51124181873101]
We aim at transforming an image with a fine-grained category to synthesize new images that preserve the identity of the input image.
We adopt a model based on generative adversarial networks to disentangle the identity related and unrelated factors of an image.
Experiments on the CompCars and Multi-PIE datasets demonstrate that our model preserves the identity of the generated images much better than the state-of-the-art image-to-image transformation models.
arXiv Detail & Related papers (2020-01-12T05:26:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.