A One-Shot Texture-Perceiving Generative Adversarial Network for
Unsupervised Surface Inspection
- URL: http://arxiv.org/abs/2106.06792v1
- Date: Sat, 12 Jun 2021 15:05:17 GMT
- Title: A One-Shot Texture-Perceiving Generative Adversarial Network for
Unsupervised Surface Inspection
- Authors: Lingyun Gu, Lin Zhang, Zhaokui Wang
- Abstract summary: We propose a hierarchical texture-perceiving generative adversarial network (HTP-GAN) that is learned from the one-shot normal image in an unsupervised scheme.
Specifically, the HTP-GAN contains a pyramid of convolutional GANs that can capture the global structure and fine-grained representation of an image simultaneously.
In the discriminator, a texture-perceiving module is devised to capture the spatially invariant representation of normal image via directional convolutions.
- Score: 4.6284467350305585
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Visual surface inspection is a challenging task owing to the highly diverse
appearance of target surfaces and defective regions. Previous attempts heavily
rely on vast quantities of training examples with manual annotation. However,
in some practical cases, it is difficult to obtain a large number of samples
for inspection. To combat it, we propose a hierarchical texture-perceiving
generative adversarial network (HTP-GAN) that is learned from the one-shot
normal image in an unsupervised scheme. Specifically, the HTP-GAN contains a
pyramid of convolutional GANs that can capture the global structure and
fine-grained representation of an image simultaneously. This innovation helps
distinguishing defective surface regions from normal ones. In addition, in the
discriminator, a texture-perceiving module is devised to capture the spatially
invariant representation of normal image via directional convolutions, making
it more sensitive to defective areas. Experiments on a variety of datasets
consistently demonstrate the effectiveness of our method.
Related papers
- LDR-Net: A Novel Framework for AI-generated Image Detection via Localized Discrepancy Representation [30.677834580640123]
We propose the localized discrepancy representation network (LDR-Net) for detecting AI-generated images.
LDR-Net captures smoothing artifacts and texture irregularities, which are common but often overlooked.
It achieves state-of-the-art performance in detecting generated images and exhibits satisfactory generalization across unseen generative models.
arXiv Detail & Related papers (2025-01-23T08:46:39Z) - Orthogonal Subspace Decomposition for Generalizable AI-Generated Image Detection [58.87142367781417]
A naively trained detector tends to favor overfitting to the limited and monotonous fake patterns, causing the feature space to become highly constrained and low-ranked.
One potential remedy is incorporating the pre-trained knowledge within the vision foundation models to expand the feature space.
By freezing the principal components and adapting only the remained components, we preserve the pre-trained knowledge while learning forgery-related patterns.
arXiv Detail & Related papers (2024-11-23T19:10:32Z) - GenFace: A Large-Scale Fine-Grained Face Forgery Benchmark and Cross Appearance-Edge Learning [50.7702397913573]
The rapid advancement of photorealistic generators has reached a critical juncture where the discrepancy between authentic and manipulated images is increasingly indistinguishable.
Although there have been a number of publicly available face forgery datasets, the forgery faces are mostly generated using GAN-based synthesis technology.
We propose a large-scale, diverse, and fine-grained high-fidelity dataset, namely GenFace, to facilitate the advancement of deepfake detection.
arXiv Detail & Related papers (2024-02-03T03:13:50Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - Towards Robust GAN-generated Image Detection: a Multi-view Completion
Representation [27.483031588071942]
GAN-generated image detection now becomes the first line of defense against the malicious uses of machine-synthesized image manipulations such as deepfakes.
We propose a robust detection framework based on a novel multi-view image completion representation.
We evaluate the generalization ability of our framework across six popular GANs at different resolutions and its robustness against a broad range of perturbation attacks.
arXiv Detail & Related papers (2023-06-02T08:38:02Z) - Parallax-Tolerant Unsupervised Deep Image Stitching [57.76737888499145]
We propose UDIS++, a parallax-tolerant unsupervised deep image stitching technique.
First, we propose a robust and flexible warp to model the image registration from global homography to local thin-plate spline motion.
To further eliminate the parallax artifacts, we propose to composite the stitched image seamlessly by unsupervised learning for seam-driven composition masks.
arXiv Detail & Related papers (2023-02-16T10:40:55Z) - Exploring Invariant Representation for Visible-Infrared Person
Re-Identification [77.06940947765406]
Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy.
In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM)
Experiment results on two standard cross-spectral person re-identification datasets, RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.
arXiv Detail & Related papers (2023-02-02T05:24:50Z) - Joint Learning of Deep Texture and High-Frequency Features for
Computer-Generated Image Detection [24.098604827919203]
We propose a joint learning strategy with deep texture and high-frequency features for CG image detection.
A semantic segmentation map is generated to guide the affine transformation operation.
The combination of the original image and the high-frequency components of the original and rendered images are fed into a multi-branch neural network equipped with attention mechanisms.
arXiv Detail & Related papers (2022-09-07T17:30:40Z) - PC-GANs: Progressive Compensation Generative Adversarial Networks for
Pan-sharpening [50.943080184828524]
We propose a novel two-step model for pan-sharpening that sharpens the MS image through the progressive compensation of the spatial and spectral information.
The whole model is composed of triple GANs, and based on the specific architecture, a joint compensation loss function is designed to enable the triple GANs to be trained simultaneously.
arXiv Detail & Related papers (2022-07-29T03:09:21Z) - HyperNet: Self-Supervised Hyperspectral Spatial-Spectral Feature
Understanding Network for Hyperspectral Change Detection [19.774857440703038]
HyperNet is a pixel-level self-supervised hyperspectral spatial-spectral understanding network.
It accomplishes pixel-wise feature representation for effective hyperspectral change detection.
Six hyperspectral datasets have been adopted to test the validity and generalization of proposed HyperNet.
arXiv Detail & Related papers (2022-07-20T03:26:03Z) - DFR: Deep Feature Reconstruction for Unsupervised Anomaly Segmentation [24.52418722578279]
This paper proposes an effective unsupervised anomaly segmentation approach.
It can detect and segment out the anomalies in small and confined regions of images.
It advances the state-of-the-art performances on several benchmark datasets.
arXiv Detail & Related papers (2020-12-13T18:30:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.