Manga Rescreening with Interpretable Screentone Representation
- URL: http://arxiv.org/abs/2306.04114v1
- Date: Wed, 7 Jun 2023 02:55:09 GMT
- Title: Manga Rescreening with Interpretable Screentone Representation
- Authors: Minshan Xie, Chengze Li, and Tien-Tsin Wong
- Abstract summary: The process of adapting or repurposing manga pages is a time-consuming task that requires manga artists to manually work on every single screentone region.
We propose an automatic manga rescreening pipeline that aims to minimize the human effort involved in manga adaptation.
Our pipeline automatically recognizes screentone regions and generates novel screentones with newly specified characteristics.
- Score: 21.638561901817866
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The process of adapting or repurposing manga pages is a time-consuming task
that requires manga artists to manually work on every single screentone region
and apply new patterns to create novel screentones across multiple panels. To
address this issue, we propose an automatic manga rescreening pipeline that
aims to minimize the human effort involved in manga adaptation. Our pipeline
automatically recognizes screentone regions and generates novel screentones
with newly specified characteristics (e.g., intensity or type). Existing manga
generation methods have limitations in understanding and synthesizing complex
tone- or intensity-varying regions. To overcome these limitations, we propose a
novel interpretable representation of screentones that disentangles their
intensity and type features, enabling better recognition and synthesis of
screentones. This interpretable screentone representation reduces ambiguity in
recognizing intensity-varying regions and provides fine-grained controls during
screentone synthesis by decoupling and anchoring the type or the intensity
feature. Our proposed method is demonstrated to be effective and convenient
through various experiments, showcasing the superiority of the newly proposed
pipeline with the interpretable screentone representations.
Related papers
- Zero-Painter: Training-Free Layout Control for Text-to-Image Synthesis [63.757624792753205]
We present Zero-Painter, a framework for layout-conditional text-to-image synthesis.
Our method utilizes object masks and individual descriptions, coupled with a global text prompt, to generate images with high fidelity.
arXiv Detail & Related papers (2024-06-06T13:02:00Z) - ReShader: View-Dependent Highlights for Single Image View-Synthesis [5.736642774848791]
We propose to split the view synthesis process into two independent tasks of pixel reshading and relocation.
During the reshading process, we take the single image as the input and adjust its shading based on the novel camera.
This reshaded image is then used as the input to an existing view synthesis method to relocate the pixels and produce the final novel view image.
arXiv Detail & Related papers (2023-09-19T15:23:52Z) - Screentone-Aware Manga Super-Resolution Using DeepLearning [3.0638744222997034]
High-quality images can hinder transmission and affect the viewing experience.
Traditional vectorization methods require a significant amount of manual parameter adjustment to process screentone.
Super-resolution can convert low-resolution images to high-resolution images while maintaining low transmission rates and providing high-quality results.
arXiv Detail & Related papers (2023-05-15T03:24:36Z) - SpaText: Spatio-Textual Representation for Controllable Image Generation [61.89548017729586]
SpaText is a new method for text-to-image generation using open-vocabulary scene control.
In addition to a global text prompt that describes the entire scene, the user provides a segmentation map.
We show its effectiveness on two state-of-the-art diffusion models: pixel-based and latent-conditional-based.
arXiv Detail & Related papers (2022-11-25T18:59:10Z) - SceneComposer: Any-Level Semantic Image Synthesis [80.55876413285587]
We propose a new framework for conditional image synthesis from semantic layouts of any precision levels.
The framework naturally reduces to text-to-image (T2I) at the lowest level with no shape information, and it becomes segmentation-to-image (S2I) at the highest level.
We introduce several novel techniques to address the challenges coming with this new setup.
arXiv Detail & Related papers (2022-11-21T18:59:05Z) - Control-NeRF: Editable Feature Volumes for Scene Rendering and
Manipulation [58.16911861917018]
We present a novel method for performing flexible, 3D-aware image content manipulation while enabling high-quality novel view synthesis.
Our model couples learnt scene-specific feature volumes with a scene agnostic neural rendering network.
We demonstrate various scene manipulations, including mixing scenes, deforming objects and inserting objects into scenes, while still producing photo-realistic results.
arXiv Detail & Related papers (2022-04-22T17:57:00Z) - DT2I: Dense Text-to-Image Generation from Region Descriptions [3.883984493622102]
We introduce dense text-to-image (DT2I) synthesis as a new task to pave the way toward more intuitive image generation.
We also propose DTC-GAN, a novel method to generate images from semantically rich region descriptions.
arXiv Detail & Related papers (2022-04-05T07:57:11Z) - Screentone-Preserved Manga Retargeting [27.415654292345355]
We propose a method that synthesizes a rescaled manga image while retaining the screentone in each screened region.
The rescaled manga shares the same region-wise screentone correspondences with the original manga, which enables us to simplify the screentone problem.
arXiv Detail & Related papers (2022-03-07T13:48:15Z) - NeLF: Practical Novel View Synthesis with Neural Light Field [93.41020940730915]
We present a practical and robust deep learning solution for the novel view synthesis of complex scenes.
In our approach, a continuous scene is represented as a light field, i.e., a set of rays, each of which has a corresponding color.
Our method achieves state-of-the-art novel view synthesis results while maintaining an interactive frame rate.
arXiv Detail & Related papers (2021-05-15T01:20:30Z) - Exploiting Aliasing for Manga Restoration [14.978972444431832]
We propose an innovative two-stage method to restore quality bitonal manga from degraded ones.
First, we predict the target resolution from the degraded manga via the Scale Estimation Network (SE-Net)
Then, at the target resolution, we restore the region-wise bitonal screentones via the Manga Restoration Network (MR-Net)
arXiv Detail & Related papers (2021-05-14T13:47:04Z) - AniGAN: Style-Guided Generative Adversarial Networks for Unsupervised
Anime Face Generation [84.52819242283852]
We propose a novel framework to translate a portrait photo-face into an anime appearance.
Our aim is to synthesize anime-faces which are style-consistent with a given reference anime-face.
Existing methods often fail to transfer the styles of reference anime-faces, or introduce noticeable artifacts/distortions in the local shapes of their generated faces.
arXiv Detail & Related papers (2021-02-24T22:47:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.