Screentone-Preserved Manga Retargeting
- URL: http://arxiv.org/abs/2203.03396v1
- Date: Mon, 7 Mar 2022 13:48:15 GMT
- Title: Screentone-Preserved Manga Retargeting
- Authors: Minshan Xie, Menghan Xia, Xueting Liu, Tien-Tsin Wong
- Abstract summary: We propose a method that synthesizes a rescaled manga image while retaining the screentone in each screened region.
The rescaled manga shares the same region-wise screentone correspondences with the original manga, which enables us to simplify the screentone problem.
- Score: 27.415654292345355
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a popular comic style, manga offers a unique impression by utilizing a
rich set of bitonal patterns, or screentones, for illustration. However,
screentones can easily be contaminated with visual-unpleasant aliasing and/or
blurriness after resampling, which harms its visualization on displays of
diverse resolutions. To address this problem, we propose the first manga
retargeting method that synthesizes a rescaled manga image while retaining the
screentone in each screened region. This is a non-trivial task as accurate
region-wise segmentation remains challenging. Fortunately, the rescaled manga
shares the same region-wise screentone correspondences with the original manga,
which enables us to simplify the screentone synthesis problem as an
anchor-based proposals selection and rearrangement problem. Specifically, we
design a novel manga sampling strategy to generate aliasing-free screentone
proposals, based on hierarchical grid-based anchors that connect the
correspondences between the original and the target rescaled manga.
Furthermore, a Recurrent Proposal Selection Module (RPSM) is proposed to
adaptively integrate these proposals for target screentone synthesis. Besides,
to deal with the translation insensitivity nature of screentones, we propose a
translation-invariant screentone loss to facilitate the training convergence.
Extensive qualitative and quantitative experiments are conducted to verify the
effectiveness of our method, and notably compelling results are achieved
compared to existing alternative techniques.
Related papers
- Layered Rendering Diffusion Model for Zero-Shot Guided Image Synthesis [60.260724486834164]
This paper introduces innovative solutions to enhance spatial controllability in diffusion models reliant on text queries.
We present two key innovations: Vision Guidance and the Layered Rendering Diffusion framework.
We apply our method to three practical applications: bounding box-to-image, semantic mask-to-image and image editing.
arXiv Detail & Related papers (2023-11-30T10:36:19Z) - Scenimefy: Learning to Craft Anime Scene via Semi-Supervised
Image-to-Image Translation [75.91455714614966]
We propose Scenimefy, a novel semi-supervised image-to-image translation framework.
Our approach guides the learning with structure-consistent pseudo paired data.
A patch-wise contrastive style loss is introduced to improve stylization and fine details.
arXiv Detail & Related papers (2023-08-24T17:59:50Z) - Manga Rescreening with Interpretable Screentone Representation [21.638561901817866]
The process of adapting or repurposing manga pages is a time-consuming task that requires manga artists to manually work on every single screentone region.
We propose an automatic manga rescreening pipeline that aims to minimize the human effort involved in manga adaptation.
Our pipeline automatically recognizes screentone regions and generates novel screentones with newly specified characteristics.
arXiv Detail & Related papers (2023-06-07T02:55:09Z) - Conditional Score Guidance for Text-Driven Image-to-Image Translation [52.73564644268749]
We present a novel algorithm for text-driven image-to-image translation based on a pretrained text-to-image diffusion model.
Our method aims to generate a target image by selectively editing the regions of interest in a source image.
arXiv Detail & Related papers (2023-05-29T10:48:34Z) - Break-A-Scene: Extracting Multiple Concepts from a Single Image [80.47666266017207]
We introduce the task of textual scene decomposition.
We propose augmenting the input image with masks that indicate the presence of target concepts.
We then present a novel two-phase customization process.
arXiv Detail & Related papers (2023-05-25T17:59:04Z) - Screentone-Aware Manga Super-Resolution Using DeepLearning [3.0638744222997034]
High-quality images can hinder transmission and affect the viewing experience.
Traditional vectorization methods require a significant amount of manual parameter adjustment to process screentone.
Super-resolution can convert low-resolution images to high-resolution images while maintaining low transmission rates and providing high-quality results.
arXiv Detail & Related papers (2023-05-15T03:24:36Z) - High-Fidelity Guided Image Synthesis with Latent Diffusion Models [50.39294302741698]
The proposed approach outperforms the previous state-of-the-art by over 85.32% on the overall user satisfaction scores.
Human user study results show that the proposed approach outperforms the previous state-of-the-art by over 85.32% on the overall user satisfaction scores.
arXiv Detail & Related papers (2022-11-30T15:43:20Z) - Exploiting Aliasing for Manga Restoration [14.978972444431832]
We propose an innovative two-stage method to restore quality bitonal manga from degraded ones.
First, we predict the target resolution from the degraded manga via the Scale Estimation Network (SE-Net)
Then, at the target resolution, we restore the region-wise bitonal screentones via the Manga Restoration Network (MR-Net)
arXiv Detail & Related papers (2021-05-14T13:47:04Z) - Semantic Layout Manipulation with High-Resolution Sparse Attention [106.59650698907953]
We tackle the problem of semantic image layout manipulation, which aims to manipulate an input image by editing its semantic label map.
A core problem of this task is how to transfer visual details from the input images to the new semantic layout while making the resulting image visually realistic.
We propose a high-resolution sparse attention module that effectively transfers visual details to new layouts at a resolution up to 512x512.
arXiv Detail & Related papers (2020-12-14T06:50:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.