Learning Graph Neural Networks for Image Style Transfer
- URL: http://arxiv.org/abs/2207.11681v1
- Date: Sun, 24 Jul 2022 07:41:31 GMT
- Title: Learning Graph Neural Networks for Image Style Transfer
- Authors: Yongcheng Jing, Yining Mao, Yiding Yang, Yibing Zhan, Mingli Song,
Xinchao Wang, Dacheng Tao
- Abstract summary: State-of-the-art parametric and non-parametric style transfer approaches are prone to either distorted local style patterns due to global statistics alignment, or unpleasing artifacts resulting from patch mismatching.
In this paper, we study a novel semi-parametric neural style transfer framework that alleviates the deficiency of both parametric and non-parametric stylization.
- Score: 131.73237185888215
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State-of-the-art parametric and non-parametric style transfer approaches are
prone to either distorted local style patterns due to global statistics
alignment, or unpleasing artifacts resulting from patch mismatching. In this
paper, we study a novel semi-parametric neural style transfer framework that
alleviates the deficiency of both parametric and non-parametric stylization.
The core idea of our approach is to establish accurate and fine-grained
content-style correspondences using graph neural networks (GNNs). To this end,
we develop an elaborated GNN model with content and style local patches as the
graph vertices. The style transfer procedure is then modeled as the
attention-based heterogeneous message passing between the style and content
nodes in a learnable manner, leading to adaptive many-to-one style-content
correlations at the local patch level. In addition, an elaborated deformable
graph convolutional operation is introduced for cross-scale style-content
matching. Experimental results demonstrate that the proposed semi-parametric
image stylization approach yields encouraging results on the challenging style
patterns, preserving both global appearance and exquisite details. Furthermore,
by controlling the number of edges at the inference stage, the proposed method
also triggers novel functionalities like diversified patch-based stylization
with a single model.
Related papers
- ZePo: Zero-Shot Portrait Stylization with Faster Sampling [61.14140480095604]
This paper presents an inversion-free portrait stylization framework based on diffusion models that accomplishes content and style feature fusion in merely four sampling steps.
We propose a feature merging strategy to amalgamate redundant features in Consistency Features, thereby reducing the computational load of attention control.
arXiv Detail & Related papers (2024-08-10T08:53:41Z) - Locally Stylized Neural Radiance Fields [30.037649804991315]
We propose a stylization framework for neural radiance fields (NeRF) based on local style transfer.
In particular, we use a hash-grid encoding to learn the embedding of the appearance and geometry components.
We show that our method yields plausible stylization results with novel view synthesis.
arXiv Detail & Related papers (2023-09-19T15:08:10Z) - WSAM: Visual Explanations from Style Augmentation as Adversarial
Attacker and Their Influence in Image Classification [2.282270386262498]
This paper outlines a style augmentation algorithm using noise-based sampling with addition to improving randomization on a general linear transformation for style transfer.
All models not only present incredible robustness against image stylizing but also outperform all previous methods and surpass the state-of-the-art performance for the STL-10 dataset.
arXiv Detail & Related papers (2023-08-29T02:50:36Z) - A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive
Learning [84.8813842101747]
Unified Contrastive Arbitrary Style Transfer (UCAST) is a novel style representation learning and transfer framework.
We present an adaptive contrastive learning scheme for style transfer by introducing an input-dependent temperature.
Our framework consists of three key components, i.e., a parallel contrastive learning scheme for style representation and style transfer, a domain enhancement module for effective learning of style distribution, and a generative network for style transfer.
arXiv Detail & Related papers (2023-03-09T04:35:00Z) - G-MSM: Unsupervised Multi-Shape Matching with Graph-based Affinity
Priors [52.646396621449]
G-MSM is a novel unsupervised learning approach for non-rigid shape correspondence.
We construct an affinity graph on a given set of training shapes in a self-supervised manner.
We demonstrate state-of-the-art performance on several recent shape correspondence benchmarks.
arXiv Detail & Related papers (2022-12-06T12:09:24Z) - Domain Enhanced Arbitrary Image Style Transfer via Contrastive Learning [84.8813842101747]
Contrastive Arbitrary Style Transfer (CAST) is a new style representation learning and style transfer method via contrastive learning.
Our framework consists of three key components, i.e., a multi-layer style projector for style code encoding, a domain enhancement module for effective learning of style distribution, and a generative network for image style transfer.
arXiv Detail & Related papers (2022-05-19T13:11:24Z) - SSR-GNNs: Stroke-based Sketch Representation with Graph Neural Networks [34.759306840182205]
This paper investigates a graph representation for sketches, where the information of strokes, i.e., parts of a sketch, are encoded on vertices and information of inter-stroke on edges.
The resultant graph representation facilitates the training of a Graph Neural Networks for classification tasks.
The proposed representation enables generation of novel sketches that are structurally similar to while separable from the existing dataset.
arXiv Detail & Related papers (2022-04-27T19:18:01Z) - StyleMeUp: Towards Style-Agnostic Sketch-Based Image Retrieval [119.03470556503942]
Crossmodal matching problem is typically solved by learning a joint embedding space where semantic content shared between photo and sketch modalities are preserved.
An effective model needs to explicitly account for this style diversity, crucially, to unseen user styles.
Our model can not only disentangle the cross-modal shared semantic content, but can adapt the disentanglement to any unseen user style as well, making the model truly agnostic.
arXiv Detail & Related papers (2021-03-29T15:44:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.