Learning Transformation-Aware Embeddings for Image Forensics
- URL: http://arxiv.org/abs/2001.04547v1
- Date: Mon, 13 Jan 2020 22:01:24 GMT
- Title: Learning Transformation-Aware Embeddings for Image Forensics
- Authors: Aparna Bharati, Daniel Moreira, Patrick Flynn, Anderson Rocha, Kevin
Bowyer, Walter Scheirer
- Abstract summary: Image Provenance Analysis aims at discovering relationships among different manipulated image versions that share content.
One of the main sub-problems for provenance analysis that has not yet been addressed directly is the edit ordering of images that share full content or are near-duplicates.
This paper introduces a novel deep learning-based approach to provide a plausible ordering to images that have been generated from a single image through transformations.
- Score: 15.484408315588569
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A dramatic rise in the flow of manipulated image content on the Internet has
led to an aggressive response from the media forensics research community. New
efforts have incorporated increased usage of techniques from computer vision
and machine learning to detect and profile the space of image manipulations.
This paper addresses Image Provenance Analysis, which aims at discovering
relationships among different manipulated image versions that share content.
One of the main sub-problems for provenance analysis that has not yet been
addressed directly is the edit ordering of images that share full content or
are near-duplicates. The existing large networks that generate image
descriptors for tasks such as object recognition may not encode the subtle
differences between these image covariates. This paper introduces a novel deep
learning-based approach to provide a plausible ordering to images that have
been generated from a single image through transformations. Our approach learns
transformation-aware descriptors using weak supervision via composited
transformations and a rank-based quadruplet loss. To establish the efficacy of
the proposed approach, comparisons with state-of-the-art handcrafted and deep
learning-based descriptors, and image matching approaches are made. Further
experimentation validates the proposed approach in the context of image
provenance analysis.
Related papers
- CricaVPR: Cross-image Correlation-aware Representation Learning for Visual Place Recognition [73.51329037954866]
We propose a robust global representation method with cross-image correlation awareness for visual place recognition.
Our method uses the attention mechanism to correlate multiple images within a batch.
Our method outperforms state-of-the-art methods by a large margin with significantly less training time.
arXiv Detail & Related papers (2024-02-29T15:05:11Z) - Improving Generalization of Image Captioning with Unsupervised Prompt
Learning [63.26197177542422]
Generalization of Image Captioning (GeneIC) learns a domain-specific prompt vector for the target domain without requiring annotated data.
GeneIC aligns visual and language modalities with a pre-trained Contrastive Language-Image Pre-Training (CLIP) model.
arXiv Detail & Related papers (2023-08-05T12:27:01Z) - Improving Human-Object Interaction Detection via Virtual Image Learning [68.56682347374422]
Human-Object Interaction (HOI) detection aims to understand the interactions between humans and objects.
In this paper, we propose to alleviate the impact of such an unbalanced distribution via Virtual Image Leaning (VIL)
A novel label-to-image approach, Multiple Steps Image Creation (MUSIC), is proposed to create a high-quality dataset that has a consistent distribution with real images.
arXiv Detail & Related papers (2023-08-04T10:28:48Z) - Masked and Adaptive Transformer for Exemplar Based Image Translation [16.93344592811513]
Cross-domain semantic matching is challenging.
We propose a masked and adaptive transformer (MAT) for learning accurate cross-domain correspondence.
We devise a novel contrastive style learning method, for acquire quality-discriminative style representations.
arXiv Detail & Related papers (2023-03-30T03:21:14Z) - Review Neural Networks about Image Transformation Based on IGC Learning
Framework with Annotated Information [13.317099281011515]
In Computer Vision (CV), many problems can be regarded as the image transformation task, e.g., semantic segmentation and style transfer.
Some surveys only review the research on style transfer or image-to-image translation, all of which are just a branch of image transformation.
This paper proposes a novel learning framework including Independent learning, Guided learning, and Cooperative learning.
arXiv Detail & Related papers (2022-06-21T07:27:47Z) - Saliency Constrained Arbitrary Image Style Transfer using SIFT and DCNN [22.57205921266602]
When common neural style transfer methods are used, the textures and colors in the style image are usually transferred imperfectly to the content image.
This paper proposes a novel saliency constrained method to reduce or avoid such effects.
The experiments show that the saliency maps of source images can help find the correct matching and avoid artifacts.
arXiv Detail & Related papers (2022-01-14T09:00:55Z) - RTIC: Residual Learning for Text and Image Composition using Graph
Convolutional Network [19.017377597937617]
We study the compositional learning of images and texts for image retrieval.
We introduce a novel method that combines the graph convolutional network (GCN) with existing composition methods.
arXiv Detail & Related papers (2021-04-07T09:41:52Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z) - Learning Spatial Relationships between Samples of Patent Image Shapes [14.37369942979269]
We propose a method suitable to binary images which bridges some of the successes of deep learning (DL)
The method consists on extracting the shape of interest from the binary image and applying a non-Euclidean geometric neural-net architecture to learn the local and global spatial relationships of the shape.
arXiv Detail & Related papers (2020-04-12T23:05:19Z) - Cross-domain Correspondence Learning for Exemplar-based Image
Translation [59.35767271091425]
We present a framework for exemplar-based image translation, which synthesizes a photo-realistic image from the input in a distinct domain.
The output has the style (e.g., color, texture) in consistency with the semantically corresponding objects in the exemplar.
We show that our method is superior to state-of-the-art methods in terms of image quality significantly.
arXiv Detail & Related papers (2020-04-12T09:10:57Z) - Fine-grained Image-to-Image Transformation towards Visual Recognition [102.51124181873101]
We aim at transforming an image with a fine-grained category to synthesize new images that preserve the identity of the input image.
We adopt a model based on generative adversarial networks to disentangle the identity related and unrelated factors of an image.
Experiments on the CompCars and Multi-PIE datasets demonstrate that our model preserves the identity of the generated images much better than the state-of-the-art image-to-image transformation models.
arXiv Detail & Related papers (2020-01-12T05:26:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.