Detection and Localization of Multiple Image Splicing Using MobileNet V1
- URL: http://arxiv.org/abs/2108.09674v1
- Date: Sun, 22 Aug 2021 09:27:22 GMT
- Title: Detection and Localization of Multiple Image Splicing Using MobileNet V1
- Authors: Kalyani Kadam, Dr. Swati Ahirrao, Dr. Ketan Kotecha, Sayan Sahu
- Abstract summary: Two or more images are combined to generate a new image that can transmit information across social media platforms.
This research work proposes multiple image splicing forgery detection using Mask R-CNN, with a backbone as a MobileNet V1.
It also calculates the percentage score of a forged region of multiple spliced images.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In modern society, digital images have become a prominent source of
information and medium of communication. They can, however, be simply altered
using freely available image editing software. Two or more images are combined
to generate a new image that can transmit information across social media
platforms to influence the people in the society. This information may have
both positive and negative consequences. Hence there is a need to develop a
technique that will detect and locates a multiple image splicing forgery in an
image. This research work proposes multiple image splicing forgery detection
using Mask R-CNN, with a backbone as a MobileNet V1. It also calculates the
percentage score of a forged region of multiple spliced images. The comparative
analysis of the proposed work with the variants of ResNet is performed. The
proposed model is trained and tested using our MISD (Multiple Image Splicing
Dataset), and it is observed that the proposed model outperforms the variants
of ResNet models (ResNet 51,101 and 151).
Related papers
- Search By Image: Deeply Exploring Beneficial Features for Beauty Product
Retrieval [21.78262478923889]
This paper studies a practically meaningful problem of beauty product retrieval (BPR) by neural networks.
We broadly extract different types of image features, and raise an intriguing question that whether these features are beneficial to i) suppress data variations of real-world captured images.
We present a novel variable-attention neural network to understand the combination of multiple features (termed VM-Net) of beauty product images.
arXiv Detail & Related papers (2023-03-24T15:38:58Z) - Impact of Scaled Image on Robustness of Deep Neural Networks [0.0]
Scaling the raw images creates out-of-distribution data, which makes it a possible adversarial attack to fool the networks.
In this work, we propose a Scaling-distortion dataset ImageNet-CS by Scaling a subset of the ImageNet Challenge dataset by different multiples.
arXiv Detail & Related papers (2022-09-02T08:06:58Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - Corrupted Image Modeling for Self-Supervised Visual Pre-Training [103.99311611776697]
We introduce Corrupted Image Modeling (CIM) for self-supervised visual pre-training.
CIM uses an auxiliary generator with a small trainable BEiT to corrupt the input image instead of using artificial mask tokens.
After pre-training, the enhancer can be used as a high-capacity visual encoder for downstream tasks.
arXiv Detail & Related papers (2022-02-07T17:59:04Z) - MVSS-Net: Multi-View Multi-Scale Supervised Networks for Image
Manipulation Detection [10.594107680952774]
Key research question for image manipulation detection is how to learn generalizable features that are sensitive to manipulations in novel data.
In this paper we address both aspects by multi-view feature learning and multi-scale supervision.
Our thoughts are realized by a new network which we term MVSS-Net and its enhanced version MVSS-Net++.
arXiv Detail & Related papers (2021-12-16T15:01:52Z) - Multimodal Deep Learning Framework for Image Popularity Prediction on
Social Media [24.104443991470287]
The popularity of an image can be affected by several factors, such as visual content, aesthetic quality, user, post metadata, and time.
We propose a deep learning model, called visual-social convolutional neural network (VSCNN), which predicts the popularity of a posted image.
arXiv Detail & Related papers (2021-05-18T19:58:58Z) - Shape-Texture Debiased Neural Network Training [50.6178024087048]
Convolutional Neural Networks are often biased towards either texture or shape, depending on the training dataset.
We develop an algorithm for shape-texture debiased learning.
Experiments show that our method successfully improves model performance on several image recognition benchmarks.
arXiv Detail & Related papers (2020-10-12T19:16:12Z) - Text as Neural Operator: Image Manipulation by Text Instruction [68.53181621741632]
In this paper, we study a setting that allows users to edit an image with multiple objects using complex text instructions to add, remove, or change the objects.
The inputs of the task are multimodal including (1) a reference image and (2) an instruction in natural language that describes desired modifications to the image.
We show that the proposed model performs favorably against recent strong baselines on three public datasets.
arXiv Detail & Related papers (2020-08-11T07:07:10Z) - Swapping Autoencoder for Deep Image Manipulation [94.33114146172606]
We propose the Swapping Autoencoder, a deep model designed specifically for image manipulation.
The key idea is to encode an image with two independent components and enforce that any swapped combination maps to a realistic image.
Experiments on multiple datasets show that our model produces better results and is substantially more efficient compared to recent generative models.
arXiv Detail & Related papers (2020-07-01T17:59:57Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.