Gradient-Free Textual Inversion
- URL: http://arxiv.org/abs/2304.05818v1
- Date: Wed, 12 Apr 2023 12:46:27 GMT
- Title: Gradient-Free Textual Inversion
- Authors: Zhengcong Fei, Mingyuan Fan, Junshi Huang
- Abstract summary: It is natural to question whether we can optimize the textual inversions by only accessing the process inference model inference model.
We introduce a emphgradient evolutionary strategy to optimize continuous textual inversion in an iterative evolutionary strategy.
Experiments in several applications demonstrate the performance of text-to-image model equipped with our proposed gradient-free method.
- Score: 34.474779413929426
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent works on personalized text-to-image generation usually learn to bind a
special token with specific subjects or styles of a few given images by tuning
its embedding through gradient descent. It is natural to question whether we
can optimize the textual inversions by only accessing the process of model
inference. As only requiring the forward computation to determine the textual
inversion retains the benefits of less GPU memory, simple deployment, and
secure access for scalable models. In this paper, we introduce a
\emph{gradient-free} framework to optimize the continuous textual inversion in
an iterative evolutionary strategy. Specifically, we first initialize an
appropriate token embedding for textual inversion with the consideration of
visual and text vocabulary information. Then, we decompose the optimization of
evolutionary strategy into dimension reduction of searching space and
non-convex gradient-free optimization in subspace, which significantly
accelerates the optimization process with negligible performance loss.
Experiments in several applications demonstrate that the performance of
text-to-image model equipped with our proposed gradient-free method is
comparable to that of gradient-based counterparts with variant GPU/CPU
platforms, flexible employment, as well as computational efficiency.
Related papers
- Efficient Personalization of Quantized Diffusion Model without Backpropagation [9.875073051988057]
Diffusion models demand extensive computational and memory resources for training, fine-tuning and inference.
We address this challenge by quantizing a diffusion model with personalization via Textual Inversion.
Our method achieves comparable performance to prior methods in image and text alignment scores for personalizing Stable Diffusion with only forward passes.
arXiv Detail & Related papers (2025-03-19T03:45:37Z) - Efficient Personalized Text-to-image Generation by Leveraging Textual Subspace [52.24866347353916]
We propose an efficient method to explore the target embedding in a textual subspace.
We also propose an efficient selection strategy for determining the basis of the textual subspace.
Our method opens the door to more efficient representation learning for personalized text-to-image generation.
arXiv Detail & Related papers (2024-06-30T06:41:21Z) - Gradient-free neural topology optimization [0.0]
gradient-free algorithms require many more iterations to converge when compared to gradient-based algorithms.
This has made them unviable for topology optimization due to the high computational cost per iteration and high dimensionality of these problems.
We propose a pre-trained neural reparameterization strategy that leads to at least one order of magnitude decrease in iteration count when optimizing the designs in latent space.
arXiv Detail & Related papers (2024-03-07T23:00:49Z) - Optimize and Reduce: A Top-Down Approach for Image Vectorization [12.998637003026273]
We propose Optimize & Reduce (O&R), a top-down approach to vectorization that is both fast and domain-agnostic.
O&R aims to attain a compact representation of input images by iteratively optimizing B'ezier curve parameters.
We demonstrate that our method is domain agnostic and outperforms existing works in both reconstruction and perceptual quality for a fixed number of shapes.
arXiv Detail & Related papers (2023-12-18T16:41:03Z) - LRANet: Towards Accurate and Efficient Scene Text Detection with
Low-Rank Approximation Network [63.554061288184165]
We propose a novel parameterized text shape method based on low-rank approximation.
By exploring the shape correlation among different text contours, our method achieves consistency, compactness, simplicity, and robustness in shape representation.
We implement an accurate and efficient arbitrary-shaped text detector named LRANet.
arXiv Detail & Related papers (2023-06-27T02:03:46Z) - A Particle-based Sparse Gaussian Process Optimizer [5.672919245950197]
We present a new swarm-swarm-based framework utilizing the underlying dynamical process of descent.
The biggest advantage of this approach is greater exploration around the current state before deciding descent descent.
arXiv Detail & Related papers (2022-11-26T09:06:15Z) - BBTv2: Pure Black-Box Optimization Can Be Comparable to Gradient Descent
for Few-Shot Learning [83.26610968655815]
Black-Box Tuning is a derivative-free approach to optimize continuous prompt tokens prepended to the input of language models.
We present BBTv2, a pure black-box optimization approach that can drive language models to achieve comparable results to gradient-based optimization.
arXiv Detail & Related papers (2022-05-23T11:10:19Z) - Gradient Descent, Stochastic Optimization, and Other Tales [8.034728173797953]
This tutorial doesn't shy away from addressing both the formal and informal aspects of gradient descent and optimization methods.
Gradient descent is one of the most popular algorithms to perform optimization and by far the most common way to optimize machine learning tasks.
In deep neural networks, the gradient followed by a single sample or a batch of samples is employed to save computational resources and escape from saddle points.
arXiv Detail & Related papers (2022-05-02T12:06:53Z) - Joint inference and input optimization in equilibrium networks [68.63726855991052]
deep equilibrium model is a class of models that foregoes traditional network depth and instead computes the output of a network by finding the fixed point of a single nonlinear layer.
We show that there is a natural synergy between these two settings.
We demonstrate this strategy on various tasks such as training generative models while optimizing over latent codes, training models for inverse problems like denoising and inpainting, adversarial training and gradient based meta-learning.
arXiv Detail & Related papers (2021-11-25T19:59:33Z) - Adapting Stepsizes by Momentumized Gradients Improves Optimization and
Generalization [89.66571637204012]
textscAdaMomentum on vision, and achieves state-the-art results consistently on other tasks including language processing.
textscAdaMomentum on vision, and achieves state-the-art results consistently on other tasks including language processing.
textscAdaMomentum on vision, and achieves state-the-art results consistently on other tasks including language processing.
arXiv Detail & Related papers (2021-06-22T03:13:23Z) - A Primer on Zeroth-Order Optimization in Signal Processing and Machine
Learning [95.85269649177336]
ZO optimization iteratively performs three major steps: gradient estimation, descent direction, and solution update.
We demonstrate promising applications of ZO optimization, such as evaluating and generating explanations from black-box deep learning models, and efficient online sensor management.
arXiv Detail & Related papers (2020-06-11T06:50:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.