CFL-Net: Image Forgery Localization Using Contrastive Learning
- URL: http://arxiv.org/abs/2210.02182v1
- Date: Tue, 4 Oct 2022 15:31:30 GMT
- Title: CFL-Net: Image Forgery Localization Using Contrastive Learning
- Authors: Fahim Faisal Niloy, Kishor Kumar Bhaumik, Simon S. Woo
- Abstract summary: We use contrastive loss to learn mapping into a feature space where the features between untampered and manipulated regions are well-separated for each image.
Our method has the advantage of localizing manipulated region without requiring any prior knowledge or assumption about the forgery type.
- Score: 16.668334854459143
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conventional forgery localizing methods usually rely on different forgery
footprints such as JPEG artifacts, edge inconsistency, camera noise, etc., with
cross-entropy loss to locate manipulated regions. However, these methods have
the disadvantage of over-fitting and focusing on only a few specific forgery
footprints. On the other hand, real-life manipulated images are generated via a
wide variety of forgery operations and thus, leave behind a wide variety of
forgery footprints. Therefore, we need a more general approach for image
forgery localization that can work well on a variety of forgery conditions. A
key assumption in underlying forged region localization is that there remains a
difference of feature distribution between untampered and manipulated regions
in each forged image sample, irrespective of the forgery type. In this paper,
we aim to leverage this difference of feature distribution to aid in image
forgery localization. Specifically, we use contrastive loss to learn mapping
into a feature space where the features between untampered and manipulated
regions are well-separated for each image. Also, our method has the advantage
of localizing manipulated region without requiring any prior knowledge or
assumption about the forgery type. We demonstrate that our work outperforms
several existing methods on three benchmark image manipulation datasets. Code
is available at https://github.com/niloy193/CFLNet.
Related papers
- TruFor: Leveraging all-round clues for trustworthy image forgery
detection and localization [17.270110456445806]
TruFor is a forensic framework that can be applied to a large variety of image manipulation methods.
We rely on the extraction of both high-level and low-level traces through a transformer-based fusion architecture.
Our method is able to reliably detect and localize both cheapfakes and deepfakes manipulations outperforming state-of-the-art works.
arXiv Detail & Related papers (2022-12-21T11:49:43Z) - PixSelect: Less but Reliable Pixels for Accurate and Efficient
Localization [0.0]
We address the problem of estimating the global 6 DoF camera pose from a single RGB image in a given environment.
Our work exceeds state-ofthe-art methods on outdoor Cambridge Landmarks dataset.
arXiv Detail & Related papers (2022-06-08T09:46:03Z) - Location-Free Camouflage Generation Network [82.74353843283407]
Camouflage is a common visual phenomenon, which refers to hiding the foreground objects into the background images, making them briefly invisible to the human eye.
This paper proposes a novel Location-free Camouflage Generation Network (LCG-Net) that fuse high-level features of foreground and background image, and generate result by one inference.
Experiments show that our method has results as satisfactory as state-of-the-art in the single-appearance regions and are less likely to be completely invisible, but far exceed the quality of the state-of-the-art in the multi-appearance regions.
arXiv Detail & Related papers (2022-03-18T10:33:40Z) - Global and Local Alignment Networks for Unpaired Image-to-Image
Translation [170.08142745705575]
The goal of unpaired image-to-image translation is to produce an output image reflecting the target domain's style.
Due to the lack of attention to the content change in existing methods, semantic information from source images suffers from degradation during translation.
We introduce a novel approach, Global and Local Alignment Networks (GLA-Net)
Our method effectively generates sharper and more realistic images than existing approaches.
arXiv Detail & Related papers (2021-11-19T18:01:54Z) - Sparse Spatial Transformers for Few-Shot Learning [6.271261279657655]
Learning from limited data is challenging because data scarcity leads to a poor generalization of the trained model.
We propose a novel transformer-based neural network architecture called sparse spatial transformers.
Our method finds task-relevant features and suppresses task-irrelevant features.
arXiv Detail & Related papers (2021-09-27T10:36:32Z) - Contrastive Unpaired Translation using Focal Loss for Patch
Classification [0.0]
Contrastive Unpaired Translation is a new method for image-to-image translation.
We show that using focal loss in place of cross-entropy loss within the PatchNCE loss can improve on the model's performance.
arXiv Detail & Related papers (2021-09-25T20:22:33Z) - Cross-Descriptor Visual Localization and Mapping [81.16435356103133]
Visual localization and mapping is the key technology underlying the majority of Mixed Reality and robotics systems.
We present three novel scenarios for localization and mapping which require the continuous update of feature representations.
Our data-driven approach is agnostic to the feature descriptor type, has low computational requirements, and scales linearly with the number of description algorithms.
arXiv Detail & Related papers (2020-12-02T18:19:51Z) - SMILE: Semantically-guided Multi-attribute Image and Layout Editing [154.69452301122175]
Attribute image manipulation has been a very active topic since the introduction of Generative Adversarial Networks (GANs)
We present a multimodal representation that handles all attributes, be it guided by random noise or images, while only using the underlying domain information of the target domain.
Our method is capable of adding, removing or changing either fine-grained or coarse attributes by using an image as a reference or by exploring the style distribution space.
arXiv Detail & Related papers (2020-10-05T20:15:21Z) - Multi-View Optimization of Local Feature Geometry [70.18863787469805]
We address the problem of refining the geometry of local image features from multiple views without known scene or camera geometry.
Our proposed method naturally complements the traditional feature extraction and matching paradigm.
We show that our method consistently improves the triangulation and camera localization performance for both hand-crafted and learned local features.
arXiv Detail & Related papers (2020-03-18T17:22:11Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.