Image Vectorization: a Review
- URL: http://arxiv.org/abs/2306.06441v1
- Date: Sat, 10 Jun 2023 13:41:02 GMT
- Title: Image Vectorization: a Review
- Authors: Maria Dziuba, Ivan Jarsky, Valeria Efimova and Andrey Filchenkov
- Abstract summary: Instead of generating vector images directly, you can first synthesize an image and then apply vectorization.
In this paper, we focus specifically on machine learning-compatible vectorization methods.
- Score: 4.258673477256579
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Nowadays, there are many diffusion and autoregressive models that show
impressive results for generating images from text and other input domains.
However, these methods are not intended for ultra-high-resolution image
synthesis. Vector graphics are devoid of this disadvantage, so the generation
of images in this format looks very promising. Instead of generating vector
images directly, you can first synthesize a raster image and then apply
vectorization. Vectorization is the process of converting a raster image into a
similar vector image using primitive shapes. Besides being similar, generated
vector image is also required to contain the minimum number of shapes for
rendering. In this paper, we focus specifically on machine learning-compatible
vectorization methods. We are considering Mang2Vec, Deep Vectorization of
Technical Drawings, DiffVG, and LIVE models. We also provide a brief overview
of existing online methods. We also recall other algorithmic methods, Im2Vec
and ClipGEN models, but they do not participate in the comparison, since there
is no open implementation of these methods or their official implementations do
not work correctly. Our research shows that despite the ability to directly
specify the number and type of shapes, existing machine learning methods work
for a very long time and do not accurately recreate the original image. We
believe that there is no fast universal automatic approach and human control is
required for every method.
Related papers
- SuperSVG: Superpixel-based Scalable Vector Graphics Synthesis [66.44553285020066]
SuperSVG is a superpixel-based vectorization model that achieves fast and high-precision image vectorization.
We propose a two-stage self-training framework, where a coarse-stage model is employed to reconstruct the main structure and a refinement-stage model is used for enriching the details.
Experiments demonstrate the superior performance of our method in terms of reconstruction accuracy and inference time compared to state-of-the-art approaches.
arXiv Detail & Related papers (2024-06-14T07:43:23Z) - Text-to-Vector Generation with Neural Path Representation [27.949704002538944]
We propose a novel neural path representation that learns the path latent space from both sequence and image modalities.
In the first stage, a pre-trained text-to-image diffusion model guides the initial generation of complex vector graphics.
In the second stage, we refine the graphics using a layer-wise image vectorization strategy to achieve clearer elements and structure.
arXiv Detail & Related papers (2024-05-16T17:59:22Z) - Unsupervised Deep Learning Image Verification Method [0.0]
The proposed method achieves a relative improvement of 56% in terms of EER over the baseline system on Labeled Faces in the Wild dataset.
This has successfully narrowed down the performance gap between cosine and PLDA scoring systems.
arXiv Detail & Related papers (2023-12-22T02:52:54Z) - Splatter Image: Ultra-Fast Single-View 3D Reconstruction [67.96212093828179]
Splatter Image is based on Gaussian Splatting, which allows fast and high-quality reconstruction of 3D scenes from multiple images.
We learn a neural network that, at test time, performs reconstruction in a feed-forward manner, at 38 FPS.
On several synthetic, real, multi-category and large-scale benchmark datasets, we achieve better results in terms of PSNR, LPIPS, and other metrics while training and evaluating much faster than prior works.
arXiv Detail & Related papers (2023-12-20T16:14:58Z) - VectorFusion: Text-to-SVG by Abstracting Pixel-Based Diffusion Models [82.93345261434943]
We show that a text-conditioned diffusion model trained on pixel representations of images can be used to generate SVG-exportable vector graphics.
Inspired by recent text-to-3D work, we learn an SVG consistent with a caption using Score Distillation Sampling.
Experiments show greater quality than prior work, and demonstrate a range of styles including pixel art and sketches.
arXiv Detail & Related papers (2022-11-21T10:04:27Z) - EpiGRAF: Rethinking training of 3D GANs [60.38818140637367]
We show that it is possible to obtain a high-resolution 3D generator with SotA image quality by following a completely different route of simply training the model patch-wise.
The resulting model, named EpiGRAF, is an efficient, high-resolution, pure 3D generator.
arXiv Detail & Related papers (2022-06-21T17:08:23Z) - Towards Layer-wise Image Vectorization [57.26058135389497]
We propose Layerwise Image Vectorization, namely LIVE, to convert images to SVGs and simultaneously maintain its image topology.
Live generates compact forms with layer-wise structures that are semantically consistent with human perspective.
Live initiates human editable SVGs for both designers and can be used in other applications.
arXiv Detail & Related papers (2022-06-09T17:55:02Z) - Im2Vec: Synthesizing Vector Graphics without Vector Supervision [31.074606918245298]
Vector graphics are widely used to represent fonts, logos, digital artworks, and graphic designs.
One can alwaysize the input graphic and resort to image-based generative approaches.
Current models that require explicit supervision on the vector representation at training time are difficult to obtain.
We propose a new neural network that can generate complex vector graphics with varying topologies.
arXiv Detail & Related papers (2021-02-04T18:39:45Z) - Swapping Autoencoder for Deep Image Manipulation [94.33114146172606]
We propose the Swapping Autoencoder, a deep model designed specifically for image manipulation.
The key idea is to encode an image with two independent components and enforce that any swapped combination maps to a realistic image.
Experiments on multiple datasets show that our model produces better results and is substantially more efficient compared to recent generative models.
arXiv Detail & Related papers (2020-07-01T17:59:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.