HairCLIP: Design Your Hair by Text and Reference Image
- URL: http://arxiv.org/abs/2112.05142v1
- Date: Thu, 9 Dec 2021 18:59:58 GMT
- Title: HairCLIP: Design Your Hair by Text and Reference Image
- Authors: Tianyi Wei and Dongdong Chen and Wenbo Zhou and Jing Liao and Zhentao
Tan and Lu Yuan and Weiming Zhang and Nenghai Yu
- Abstract summary: This paper proposes a new hair editing interaction mode, which enables manipulating hair attributes individually or jointly.
We encode the image and text conditions in a shared embedding space and propose a unified hair editing framework.
With the carefully designed network structures and loss functions, our framework can perform high-quality hair editing.
- Score: 100.85116679883724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hair editing is an interesting and challenging problem in computer vision and
graphics. Many existing methods require well-drawn sketches or masks as
conditional inputs for editing, however these interactions are neither
straightforward nor efficient. In order to free users from the tedious
interaction process, this paper proposes a new hair editing interaction mode,
which enables manipulating hair attributes individually or jointly based on the
texts or reference images provided by users. For this purpose, we encode the
image and text conditions in a shared embedding space and propose a unified
hair editing framework by leveraging the powerful image text representation
capability of the Contrastive Language-Image Pre-Training (CLIP) model. With
the carefully designed network structures and loss functions, our framework can
perform high-quality hair editing in a disentangled manner. Extensive
experiments demonstrate the superiority of our approach in terms of
manipulation accuracy, visual realism of editing results, and irrelevant
attribute preservation. Project repo is https://github.com/wty-ustc/HairCLIP.
Related papers
- Text-Driven Image Editing via Learnable Regions [74.45313434129005]
We introduce a method for region-based image editing driven by textual prompts, without the need for user-provided masks or sketches.
We show that this simple approach enables flexible editing that is compatible with current image generation models.
Experiments demonstrate the competitive performance of our method in manipulating images with high fidelity and realism that correspond to the provided language descriptions.
arXiv Detail & Related papers (2023-11-28T02:27:31Z) - HairCLIPv2: Unifying Hair Editing via Proxy Feature Blending [94.39417893972262]
HairCLIP is the first work that enables hair editing based on text descriptions or reference images.
In this paper, we propose HairCLIPv2, aiming to support all the aforementioned interactions with one unified framework.
The key idea is to convert all the hair editing tasks into hair transfer tasks, with editing conditions converted into different proxies accordingly.
arXiv Detail & Related papers (2023-10-16T17:59:58Z) - Visual Instruction Inversion: Image Editing via Visual Prompting [34.96778567507126]
We present a method for image editing via visual prompting.
We leverage the rich, pretrained editing capabilities of text-to-image diffusion models by inverting visual prompts into editing instructions.
arXiv Detail & Related papers (2023-07-26T17:50:10Z) - CLIP-Guided StyleGAN Inversion for Text-Driven Real Image Editing [22.40686064568406]
We present CLIPInverter, a new text-driven image editing approach that is able to efficiently and reliably perform multi-attribute changes.
Our method outperforms competing approaches in terms of manipulation accuracy and photo-realism on various domains including human faces, cats, and birds.
arXiv Detail & Related papers (2023-07-17T11:29:48Z) - FICE: Text-Conditioned Fashion Image Editing With Guided GAN Inversion [16.583537785874604]
We propose a novel text-conditioned editing model, called FICE, capable of handling a wide variety of diverse text descriptions.
FICE generates highly realistic fashion images and leads to stronger editing performance than existing competing approaches.
arXiv Detail & Related papers (2023-01-05T15:33:23Z) - Imagen Editor and EditBench: Advancing and Evaluating Text-Guided Image
Inpainting [53.708523312636096]
We present Imagen Editor, a cascaded diffusion model built, by fine-tuning on text-guided image inpainting.
edits are faithful to the text prompts, which is accomplished by using object detectors to propose inpainting masks during training.
To improve qualitative and quantitative evaluation, we introduce EditBench, a systematic benchmark for text-guided image inpainting.
arXiv Detail & Related papers (2022-12-13T21:25:11Z) - Towards Counterfactual Image Manipulation via CLIP [106.94502632502194]
Existing methods can achieve realistic editing of different visual attributes such as age and gender of facial images.
We investigate this problem in a text-driven manner with Contrastive-Language-Image-Pretraining (CLIP)
We design a novel contrastive loss that exploits predefined CLIP-space directions to guide the editing toward desired directions from different perspectives.
arXiv Detail & Related papers (2022-07-06T17:02:25Z) - MichiGAN: Multi-Input-Conditioned Hair Image Generation for Portrait
Editing [122.82964863607938]
MichiGAN is a novel conditional image generation method for interactive portrait hair manipulation.
We provide user control over every major hair visual factor, including shape, structure, appearance, and background.
We also build an interactive portrait hair editing system that enables straightforward manipulation of hair by projecting intuitive and high-level user inputs.
arXiv Detail & Related papers (2020-10-30T17:59:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.