Transformer-based Graph Neural Networks for Outfit Generation
- URL: http://arxiv.org/abs/2304.08098v1
- Date: Mon, 17 Apr 2023 09:18:45 GMT
- Title: Transformer-based Graph Neural Networks for Outfit Generation
- Authors: Federico Becattini, Federico Maria Teotini, Alberto Del Bimbo
- Abstract summary: TGNN exploits multi-headed self attention to capture relations between clothing items in a graph as a message passing step in Convolutional Graph Neural Networks.
We propose a transformer-based architecture, which exploits multi-headed self attention to capture relations between clothing items in a graph as a message passing step in Convolutional Graph Neural Networks.
- Score: 22.86041284499166
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Suggesting complementary clothing items to compose an outfit is a process of
emerging interest, yet it involves a fine understanding of fashion trends and
visual aesthetics. Previous works have mainly focused on recommendation by
scoring visual appeal and representing garments as ordered sequences or as
collections of pairwise-compatible items. This limits the full usage of
relations among clothes. We attempt to bridge the gap between outfit
recommendation and generation by leveraging a graph-based representation of
items in a collection. The work carried out in this paper, tries to build a
bridge between outfit recommendation and generation, by discovering new
appealing outfits starting from a collection of pre-existing ones. We propose a
transformer-based architecture, named TGNN, which exploits multi-headed self
attention to capture relations between clothing items in a graph as a message
passing step in Convolutional Graph Neural Networks. Specifically, starting
from a seed, i.e.~one or more garments, outfit generation is performed by
iteratively choosing the garment that is most compatible with the previously
chosen ones. Extensive experimentations are conducted with two different
datasets, demonstrating the capability of the model to perform seeded outfit
generation as well as obtaining state of the art results on compatibility
estimation tasks.
Related papers
- Multi-Garment Customized Model Generation [3.1679243514285194]
Multi-Garment Customized Model Generation is a unified framework based on Latent Diffusion Models (LDMs)
Our framework supports the conditional generation of multiple garments through decoupled multi-garment feature fusion.
The proposed garment encoder is a plug-and-play module that can be combined with other extension modules.
arXiv Detail & Related papers (2024-08-09T17:57:33Z) - IMAGDressing-v1: Customizable Virtual Dressing [58.44155202253754]
IMAGDressing-v1 is a virtual dressing task that generates freely editable human images with fixed garments and optional conditions.
IMAGDressing-v1 incorporates a garment UNet that captures semantic features from CLIP and texture features from VAE.
We present a hybrid attention module, including a frozen self-attention and a trainable cross-attention, to integrate garment features from the garment UNet into a frozen denoising UNet.
arXiv Detail & Related papers (2024-07-17T16:26:30Z) - Fashion Recommendation: Outfit Compatibility using GNN [0.0]
We follow two existing approaches that employ graphs to represent outfits.
Both Node-wise Graph Neural Network (NGNN) and Hypergraph Neural Network aim to score a set of items according to the outfit compatibility of items.
We recreate the analysis on a subset of this data and compare the two existing models on their performance on two tasks Fill in the blank (FITB): finding an item that completes an outfit, and Compatibility prediction: estimating compatibility of different items grouped as an outfit.
arXiv Detail & Related papers (2024-04-28T00:57:17Z) - VICTOR: Visual Incompatibility Detection with Transformers and
Fashion-specific contrastive pre-training [18.753508811614644]
Visual InCompatibility TransfORmer (VICTOR) is optimized for two tasks: 1) overall compatibility as regression and 2) the detection of mismatching items.
We build upon the Polyvore outfit benchmark to generate partially mismatching outfits, creating a new dataset termed Polyvore-MISFITs.
A series of ablation and comparative analyses show that the proposed architecture can compete and even surpass the current state-of-the-art on Polyvore datasets.
arXiv Detail & Related papers (2022-07-27T11:18:55Z) - Learning Fashion Compatibility from In-the-wild Images [6.591937706757015]
We propose to learn representations for compatibility prediction from in-the-wild street fashion images through self-supervised learning.
Our pretext task is formulated such that the representations of different items worn by the same person are closer compared to those worn by other people.
We conduct experiments on two popular fashion compatibility benchmarks - Polyvore and Polyvore-Disjoint outfits.
arXiv Detail & Related papers (2022-06-13T09:05:25Z) - Clothes-Changing Person Re-identification with RGB Modality Only [102.44387094119165]
We propose a Clothes-based Adrial Loss (CAL) to mine clothes-irrelevant features from the original RGB images.
Videos contain richer appearance and additional temporal information, which can be used to model propertemporal patterns.
arXiv Detail & Related papers (2022-04-14T11:38:28Z) - UIGR: Unified Interactive Garment Retrieval [105.56179829647142]
Interactive garment retrieval (IGR) aims to retrieve a target garment image based on a reference garment image.
Two IGR tasks have been studied extensively: text-guided garment retrieval (TGR) and visually compatible garment retrieval (VCR)
We propose a Unified Interactive Garment Retrieval (UIGR) framework to unify TGR and VCR.
arXiv Detail & Related papers (2022-04-06T21:54:14Z) - Arbitrary Virtual Try-On Network: Characteristics Preservation and
Trade-off between Body and Clothing [85.74977256940855]
We propose an Arbitrary Virtual Try-On Network (AVTON) for all-type clothes.
AVTON can synthesize realistic try-on images by preserving and trading off characteristics of the target clothes and the reference person.
Our approach can achieve better performance compared with the state-of-the-art virtual try-on methods.
arXiv Detail & Related papers (2021-11-24T08:59:56Z) - Apparel-invariant Feature Learning for Apparel-changed Person
Re-identification [70.16040194572406]
Most public ReID datasets are collected in a short time window in which persons' appearance rarely changes.
In real-world applications such as in a shopping mall, the same person's clothing may change, and different persons may wearing similar clothes.
It is critical to learn an apparel-invariant person representation under cases like cloth changing or several persons wearing similar clothes.
arXiv Detail & Related papers (2020-08-14T03:49:14Z) - Fashion Recommendation and Compatibility Prediction Using Relational
Network [18.13692056232815]
We develop a Relation Network (RN) to develop new compatibility learning models.
FashionRN learns the compatibility of an entire outfit, with an arbitrary number of items, in an arbitrary order.
We evaluate our model using a large dataset of 49,740 outfits that we collected from Polyvore website.
arXiv Detail & Related papers (2020-05-13T21:00:54Z) - Learning Diverse Fashion Collocation by Neural Graph Filtering [78.9188246136867]
We propose a novel fashion collocation framework, Neural Graph Filtering, that models a flexible set of fashion items via a graph neural network.
By applying symmetric operations on the edge vectors, this framework allows varying numbers of inputs/outputs and is invariant to their ordering.
We evaluate the proposed approach on three popular benchmarks, the Polyvore dataset, the Polyvore-D dataset, and our reorganized Amazon Fashion dataset.
arXiv Detail & Related papers (2020-03-11T16:17:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.