OmniVTON++: Training-Free Universal Virtual Try-On with Principal Pose Guidance
- URL: http://arxiv.org/abs/2602.14552v1
- Date: Mon, 16 Feb 2026 08:27:43 GMT
- Title: OmniVTON++: Training-Free Universal Virtual Try-On with Principal Pose Guidance
- Authors: Zhaotong Yang, Yong Du, Shengfeng He, Yuhui Li, Xinzhe Li, Yangyang Xu, Junyu Dong, Jian Yang,
- Abstract summary: Image-based Virtual Try-On (VTON) concerns the synthesis of realistic person imagery through garment re-rendering under human pose and body constraints.<n>We present OmniVTON++, a training-free VTON framework designed for universal applicability.
- Score: 85.23143742905695
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image-based Virtual Try-On (VTON) concerns the synthesis of realistic person imagery through garment re-rendering under human pose and body constraints. In practice, however, existing approaches are typically optimized for specific data conditions, making their deployment reliant on retraining and limiting their generalization as a unified solution. We present OmniVTON++, a training-free VTON framework designed for universal applicability. It addresses the intertwined challenges of garment alignment, human structural coherence, and boundary continuity by coordinating Structured Garment Morphing for correspondence-driven garment adaptation, Principal Pose Guidance for step-wise structural regulation during diffusion sampling, and Continuous Boundary Stitching for boundary-aware refinement, forming a cohesive pipeline without task-specific retraining. Experimental results demonstrate that OmniVTON++ achieves state-of-the-art performance across diverse generalization settings, including cross-dataset and cross-garment-type evaluations, while reliably operating across scenarios and diffusion backbones within a single formulation. In addition to single-garment, single-human cases, the framework supports multi-garment, multi-human, and anime character virtual try-on, expanding the scope of virtual try-on applications. The source code will be released to the public.
Related papers
- MuGa-VTON: Multi-Garment Virtual Try-On via Diffusion Transformers with Prompt Customization [19.780800887427937]
We introduce MuGa-VTON, a unified multi-garment diffusion framework that jointly models upper and lower garments together with person identity in a shared latent space.<n>This architecture supports prompt-based customization, allowing fine-grained garment modifications with minimal user input.
arXiv Detail & Related papers (2025-08-11T21:45:07Z) - Voost: A Unified and Scalable Diffusion Transformer for Bidirectional Virtual Try-On and Try-Off [7.792838907882765]
We propose Voost - a unified framework that jointly learns virtual try-on and try-off with a single diffusion transformer.<n>Voost achieves state-of-the-art results on both try-on and try-off benchmarks, consistently outperforming strong baselines in alignment accuracy, visual fidelity, and generalization.
arXiv Detail & Related papers (2025-08-06T19:10:58Z) - OmniVTON: Training-Free Universal Virtual Try-On [53.31945401098557]
Image-based Virtual Try-On (VTON) techniques rely on either supervised in-shop approaches, or unsupervised in-the-wild methods, which improve adaptability but remain constrained by data biases and limited universality.<n>We propose OmniVTON, the first training-free universal VTON framework that decouples garment and pose conditioning to achieve both texture fidelity and pose consistency across diverse settings.
arXiv Detail & Related papers (2025-07-20T16:37:53Z) - CrossVTON: Mimicking the Logic Reasoning on Cross-category Virtual Try-on guided by Tri-zone Priors [63.95051258676488]
CrossVTON is a framework for generating robust fitting images for cross-category virtual try-on.<n>It disentangles the complex reasoning required for cross-category try-on into a structured framework.<n>It achieves state-of-the-art performance, surpassing existing baselines in both qualitative and quantitative evaluations.
arXiv Detail & Related papers (2025-02-20T09:05:35Z) - ITVTON: Virtual Try-On Diffusion Transformer Based on Integrated Image and Text [1.7071356210178177]
ITVTON is an efficient framework that leverages the Diffusion Transformer (DiT) as its single generator to improve image fidelity.<n>ITVTON effectively captures garment and person images along the width dimension and incorporating textual descriptions from both.<n>Experiments on 10,257 image pairs from IGPair confirm ITVTON's robustness in real-world scenarios.
arXiv Detail & Related papers (2025-01-28T07:24:15Z) - GP-VTON: Towards General Purpose Virtual Try-on via Collaborative
Local-Flow Global-Parsing Learning [63.8668179362151]
Virtual Try-ON aims to transfer an in-shop garment onto a specific person.
Existing methods employ a global warping module to model the anisotropic deformation for different garment parts.
We propose an innovative Local-Flow Global-Parsing (LFGP) warping module and a Dynamic Gradient Truncation (DGT) training strategy.
arXiv Detail & Related papers (2023-03-24T02:12:29Z) - PASTA-GAN++: A Versatile Framework for High-Resolution Unpaired Virtual
Try-on [70.12285433529998]
PASTA-GAN++ is a versatile system for high-resolution unpaired virtual try-on.
It supports unsupervised training, arbitrary garment categories, and controllable garment editing.
arXiv Detail & Related papers (2022-07-27T11:47:49Z) - Towards Scalable Unpaired Virtual Try-On via Patch-Routed
Spatially-Adaptive GAN [66.3650689395967]
We propose a texture-preserving end-to-end network, the PAtch-routed SpaTially-Adaptive GAN (PASTA-GAN), that facilitates real-world unpaired virtual try-on.
To disentangle the style and spatial information of each garment, PASTA-GAN consists of an innovative patch-routed disentanglement module.
arXiv Detail & Related papers (2021-11-20T08:36:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.