Learning Intrinsic Images for Clothing
- URL: http://arxiv.org/abs/2111.08521v1
- Date: Tue, 16 Nov 2021 14:43:12 GMT
- Title: Learning Intrinsic Images for Clothing
- Authors: Kuo Jiang, Zian Wang, Xiaodong Yang
- Abstract summary: In this paper, we focus on intrinsic image decomposition for clothing images.
A more interpretable edge-aware metric and an annotation scheme is designed for the testing set.
We show that our proposed model significantly reduce texture-copying artifacts while retaining surprisingly tiny details.
- Score: 10.21096394185778
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Reconstruction of human clothing is an important task and often relies on
intrinsic image decomposition. With a lack of domain-specific data and coarse
evaluation metrics, existing models failed to produce satisfying results for
graphics applications. In this paper, we focus on intrinsic image decomposition
for clothing images and have comprehensive improvements. We collected
CloIntrinsics, a clothing intrinsic image dataset, including a synthetic
training set and a real-world testing set. A more interpretable edge-aware
metric and an annotation scheme is designed for the testing set, which allows
diagnostic evaluation for intrinsic models. Finally, we propose ClothInNet
model with carefully designed loss terms and an adversarial module. It utilizes
easy-to-acquire labels to learn from real-world shading, significantly improves
performance with only minor additional annotation effort. We show that our
proposed model significantly reduce texture-copying artifacts while retaining
surprisingly tiny details, outperforming existing state-of-the-art methods.
Related papers
- Diverse and Tailored Image Generation for Zero-shot Multi-label Classification [3.354528906571718]
zero-shot multi-label classification has garnered considerable attention for its capacity to operate predictions on unseen labels without human annotations.
prevailing approaches often use seen classes as imperfect proxies for unseen ones, resulting in suboptimal performance.
We propose an innovative solution: generating synthetic data to construct a training set explicitly tailored for proxyless training on unseen labels.
arXiv Detail & Related papers (2024-04-04T01:34:36Z) - Scaling Laws of Synthetic Images for Model Training ... for Now [54.43596959598466]
We study the scaling laws of synthetic images generated by state of the art text-to-image models.
We observe that synthetic images demonstrate a scaling trend similar to, but slightly less effective than, real images in CLIP training.
arXiv Detail & Related papers (2023-12-07T18:59:59Z) - Perceptual Artifacts Localization for Image Synthesis Tasks [59.638307505334076]
We introduce a novel dataset comprising 10,168 generated images, each annotated with per-pixel perceptual artifact labels.
A segmentation model, trained on our proposed dataset, effectively localizes artifacts across a range of tasks.
We propose an innovative zoom-in inpainting pipeline that seamlessly rectifies perceptual artifacts in the generated images.
arXiv Detail & Related papers (2023-10-09T10:22:08Z) - Evaluating Data Attribution for Text-to-Image Models [62.844382063780365]
We evaluate attribution through "customization" methods, which tune an existing large-scale model toward a given exemplar object or style.
Our key insight is that this allows us to efficiently create synthetic images that are computationally influenced by the exemplar by construction.
By taking into account the inherent uncertainty of the problem, we can assign soft attribution scores over a set of training images.
arXiv Detail & Related papers (2023-06-15T17:59:51Z) - TexPose: Neural Texture Learning for Self-Supervised 6D Object Pose
Estimation [55.94900327396771]
We introduce neural texture learning for 6D object pose estimation from synthetic data.
We learn to predict realistic texture of objects from real image collections.
We learn pose estimation from pixel-perfect synthetic data.
arXiv Detail & Related papers (2022-12-25T13:36:32Z) - Learning to Annotate Part Segmentation with Gradient Matching [58.100715754135685]
This paper focuses on tackling semi-supervised part segmentation tasks by generating high-quality images with a pre-trained GAN.
In particular, we formulate the annotator learning as a learning-to-learn problem.
We show that our method can learn annotators from a broad range of labelled images including real images, generated images, and even analytically rendered images.
arXiv Detail & Related papers (2022-11-06T01:29:22Z) - Understanding invariance via feedforward inversion of discriminatively
trained classifiers [30.23199531528357]
Past research has discovered that some extraneous visual detail remains in the output logits.
We develop a feedforward inversion model that produces remarkably high fidelity reconstructions.
Our approach is based on BigGAN, with conditioning on logits instead of one-hot class labels.
arXiv Detail & Related papers (2021-03-15T17:56:06Z) - From ImageNet to Image Classification: Contextualizing Progress on
Benchmarks [99.19183528305598]
We study how specific design choices in the ImageNet creation process impact the fidelity of the resulting dataset.
Our analysis pinpoints how a noisy data collection pipeline can lead to a systematic misalignment between the resulting benchmark and the real-world task it serves as a proxy for.
arXiv Detail & Related papers (2020-05-22T17:39:16Z) - TailorGAN: Making User-Defined Fashion Designs [28.805686791183618]
We propose a novel self-supervised model to synthesize garment images with disentangled attributes without paired data.
Our method consists of a reconstruction learning step and an adversarial learning step.
Experiments on this dataset and real-world samples demonstrate that our method can synthesize much better results than the state-of-the-art methods.
arXiv Detail & Related papers (2020-01-17T16:54:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.