Optimize and Reduce: A Top-Down Approach for Image Vectorization
- URL: http://arxiv.org/abs/2312.11334v1
- Date: Mon, 18 Dec 2023 16:41:03 GMT
- Title: Optimize and Reduce: A Top-Down Approach for Image Vectorization
- Authors: Or Hirschorn, Amir Jevnisek, Shai Avidan
- Abstract summary: We propose Optimize & Reduce (O&R), a top-down approach to vectorization that is both fast and domain-agnostic.
O&R aims to attain a compact representation of input images by iteratively optimizing B'ezier curve parameters.
We demonstrate that our method is domain agnostic and outperforms existing works in both reconstruction and perceptual quality for a fixed number of shapes.
- Score: 12.998637003026273
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vector image representation is a popular choice when editability and
flexibility in resolution are desired. However, most images are only available
in raster form, making raster-to-vector image conversion (vectorization) an
important task. Classical methods for vectorization are either domain-specific
or yield an abundance of shapes which limits editability and interpretability.
Learning-based methods, that use differentiable rendering, have revolutionized
vectorization, at the cost of poor generalization to out-of-training
distribution domains, and optimization-based counterparts are either slow or
produce non-editable and redundant shapes. In this work, we propose Optimize &
Reduce (O&R), a top-down approach to vectorization that is both fast and
domain-agnostic. O&R aims to attain a compact representation of input images by
iteratively optimizing B\'ezier curve parameters and significantly reducing the
number of shapes, using a devised importance measure. We contribute a benchmark
of five datasets comprising images from a broad spectrum of image complexities
- from emojis to natural-like images. Through extensive experiments on hundreds
of images, we demonstrate that our method is domain agnostic and outperforms
existing works in both reconstruction and perceptual quality for a fixed number
of shapes. Moreover, we show that our algorithm is $\times 10$ faster than the
state-of-the-art optimization-based method.
Related papers
- SuperSVG: Superpixel-based Scalable Vector Graphics Synthesis [66.44553285020066]
SuperSVG is a superpixel-based vectorization model that achieves fast and high-precision image vectorization.
We propose a two-stage self-training framework, where a coarse-stage model is employed to reconstruct the main structure and a refinement-stage model is used for enriching the details.
Experiments demonstrate the superior performance of our method in terms of reconstruction accuracy and inference time compared to state-of-the-art approaches.
arXiv Detail & Related papers (2024-06-14T07:43:23Z) - Layered Image Vectorization via Semantic Simplification [46.23779847614095]
This work presents a novel progressive image vectorization technique aimed at generating layered vectors that represent the original image from coarse to fine detail levels.
Our approach introduces semantic simplification, which combines Score Distillation Sampling and semantic segmentation to iteratively simplify the input image.
Our method provides robust optimization, which avoids local minima and enables adjustable detail levels in the final output.
arXiv Detail & Related papers (2024-06-08T08:54:35Z) - Effective Invertible Arbitrary Image Rescaling [77.46732646918936]
Invertible Neural Networks (INN) are able to increase upscaling accuracy significantly by optimizing the downscaling and upscaling cycle jointly.
A simple and effective invertible arbitrary rescaling network (IARN) is proposed to achieve arbitrary image rescaling by training only one model in this work.
It is shown to achieve a state-of-the-art (SOTA) performance in bidirectional arbitrary rescaling without compromising perceptual quality in LR outputs.
arXiv Detail & Related papers (2022-09-26T22:22:30Z) - A training-free recursive multiresolution framework for diffeomorphic
deformable image registration [6.929709872589039]
We propose a novel diffeomorphic training-free approach for deformable image registration.
The proposed architecture is simple in design. The moving image is warped successively at each resolution and finally aligned to the fixed image.
The entire system is end-to-end and optimized for each pair of images from scratch.
arXiv Detail & Related papers (2022-02-01T15:17:17Z) - Differentiable Rendering with Perturbed Optimizers [85.66675707599782]
Reasoning about 3D scenes from their 2D image projections is one of the core problems in computer vision.
Our work highlights the link between some well-known differentiable formulations and randomly smoothed renderings.
We apply our method to 3D scene reconstruction and demonstrate its advantages on the tasks of 6D pose estimation and 3D mesh reconstruction.
arXiv Detail & Related papers (2021-10-18T08:56:23Z) - Spatially-Adaptive Pixelwise Networks for Fast Image Translation [57.359250882770525]
We introduce a new generator architecture, aimed at fast and efficient high-resolution image-to-image translation.
We use pixel-wise networks; that is, each pixel is processed independently of others.
Our model is up to 18x faster than state-of-the-art baselines.
arXiv Detail & Related papers (2020-12-05T10:02:03Z) - A Flexible Framework for Designing Trainable Priors with Adaptive
Smoothing and Game Encoding [57.1077544780653]
We introduce a general framework for designing and training neural network layers whose forward passes can be interpreted as solving non-smooth convex optimization problems.
We focus on convex games, solved by local agents represented by the nodes of a graph and interacting through regularization functions.
This approach is appealing for solving imaging problems, as it allows the use of classical image priors within deep models that are trainable end to end.
arXiv Detail & Related papers (2020-06-26T08:34:54Z) - Transforming and Projecting Images into Class-conditional Generative
Networks [44.79971598515697]
We present a method for projecting an input image into the space of a class-conditional generative neural network.
Specifically, we demonstrate that one can solve for image translation, scale, and global color transformation.
We show the effectiveness of our method on real images and further demonstrate how the corresponding projections lead to better editability of these images.
arXiv Detail & Related papers (2020-05-04T17:57:47Z) - Learning Deformable Image Registration from Optimization: Perspective,
Modules, Bilevel Training and Beyond [62.730497582218284]
We develop a new deep learning based framework to optimize a diffeomorphic model via multi-scale propagation.
We conduct two groups of image registration experiments on 3D volume datasets including image-to-atlas registration on brain MRI data and image-to-image registration on liver CT data.
arXiv Detail & Related papers (2020-04-30T03:23:45Z) - Image Denoising Using Sparsifying Transform Learning and Weighted
Singular Values Minimization [7.472473280743767]
In image denoising (IDN) processing, the low-rank property is usually considered as an important image prior.
As a convex relaxation approximation of low rank, nuclear norm based algorithms and their variants have attracted significant attention.
By taking both advantages of image domain minimization and transform domain in a general framework, we propose a sparsity learning transform method.
arXiv Detail & Related papers (2020-04-02T00:30:29Z) - Optimized Feature Space Learning for Generating Efficient Binary Codes
for Image Retrieval [9.470008343329892]
We propose an approach for learning low dimensional optimized feature space with minimum intra-class variance and maximum inter-class variance.
We binarize our generated feature vectors with the popular Iterative Quantization (ITQ) approach and also propose an ensemble network to generate binary codes of desired bit length for image retrieval.
arXiv Detail & Related papers (2020-01-30T15:30:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.