AesRec: A Dataset for Aesthetics-Aligned Clothing Outfit Recommendation
- URL: http://arxiv.org/abs/2602.03416v1
- Date: Tue, 03 Feb 2026 11:44:00 GMT
- Title: AesRec: A Dataset for Aesthetics-Aligned Clothing Outfit Recommendation
- Authors: Wenxin Ye, Lin Li, Ming Li, Yang Shen, Kanghong Wang, Jimmy Xiangji Huang,
- Abstract summary: We present the AesRec benchmark dataset featuring systematic quantitative aesthetic annotations.<n>At the item level, six dimensions are independently assessed: silhouette, chromaticity, materiality, craftsmanship, wearability, and item-level impression.<n>We conduct rigorous human-machine consistency validation on a fashion dataset, confirming the reliability of the generated ratings.
- Score: 17.478482513222826
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Clothing recommendation extends beyond merely generating personalized outfits; it serves as a crucial medium for aesthetic guidance. However, existing methods predominantly rely on user-item-outfit interaction behaviors while overlooking explicit representations of clothing aesthetics. To bridge this gap, we present the AesRec benchmark dataset featuring systematic quantitative aesthetic annotations, thereby enabling the development of aesthetics-aligned recommendation systems. Grounded in professional apparel quality standards and fashion aesthetic principles, we define a multidimensional set of indicators. At the item level, six dimensions are independently assessed: silhouette, chromaticity, materiality, craftsmanship, wearability, and item-level impression. Transitioning to the outfit level, the evaluation retains the first five core attributes while introducing stylistic synergy, visual harmony, and outfit-level impression as distinct metrics to capture the collective aesthetic impact. Given the increasing human-like proficiency of Vision-Language Models in multimodal understanding and interaction, we leverage them for large-scale aesthetic scoring. We conduct rigorous human-machine consistency validation on a fashion dataset, confirming the reliability of the generated ratings. Experimental results based on AesRec further demonstrate that integrating quantified aesthetic information into clothing recommendation models can provide aesthetic guidance for users while fulfilling their personalized requirements.
Related papers
- TATTOO: Training-free AesTheTic-aware Outfit recOmmendation [9.087314807392415]
TATTOO is a Training-free AesTheTic-aware Outfit recommendation approach.<n>It first generates a target-item description using MLLMs, followed by an aesthetic chain-of-thought used to distill the images into a structured aesthetic profile.<n>Experiments on a real-world evaluation set Aesthetic-100 show that TATTOO achieves state-of-the-art performance compared with existing training-based methods.
arXiv Detail & Related papers (2025-09-27T10:46:55Z) - StyleTailor: Towards Personalized Fashion Styling via Hierarchical Negative Feedback [11.510316659758718]
StyleTailor is the first collaborative agent framework that unifies personalized apparel design, shopping recommendation, virtual try-on, and systematic evaluation into a cohesive workflow.<n>Our framework features two core agents, i.e., Designer for personalized garment selection and Consultant for virtual try-on, whose outputs are progressively refined via hierarchical vision-language model feedback.<n>To assess the performance, we introduce a comprehensive evaluation suite encompassing style consistency, visual quality, face similarity, and artistic appraisal.
arXiv Detail & Related papers (2025-08-06T08:42:42Z) - Learning to Synthesize Compatible Fashion Items Using Semantic Alignment and Collocation Classification: An Outfit Generation Framework [59.09707044733695]
We propose a novel outfit generation framework, i.e., OutfitGAN, with the aim of synthesizing an entire outfit.<n> OutfitGAN includes a semantic alignment module, which is responsible for characterizing the mapping correspondence between the existing fashion items and the synthesized ones.<n>In order to evaluate the performance of our proposed models, we built a large-scale dataset consisting of 20,000 fashion outfits.
arXiv Detail & Related papers (2025-02-05T12:13:53Z) - Prompt2Fashion: An automatically generated fashion dataset [1.8817715864806608]
We leverage generative models to automatically construct a fashion image dataset tailored to various occasions, styles, and body types as instructed by users.
We use different Large Language Models (LLMs) and prompting strategies to offer personalized outfits of high aesthetic quality, detail, and relevance to both expert and non-expert users' requirements.
arXiv Detail & Related papers (2024-09-10T11:48:05Z) - Aligning Vision Models with Human Aesthetics in Retrieval: Benchmarks and Algorithms [91.19304518033144]
We aim to align vision models with human aesthetic standards in a retrieval system.
We propose a preference-based reinforcement learning method that fine-tunes the vision models to better align the vision models with human aesthetics.
arXiv Detail & Related papers (2024-06-13T17:59:20Z) - Content and Salient Semantics Collaboration for Cloth-Changing Person Re-Identification [74.10897798660314]
Cloth-changing person re-identification aims at recognizing the same person with clothing changes across non-overlapping cameras.<n>We propose a unified Semantics Mining and Refinement (SMR) module to extract robust identity-related content and salient semantics, mitigating interference from clothing appearances effectively.<n>Our proposed method achieves state-of-the-art performance on three cloth-changing benchmarks, demonstrating its superiority over advanced competitors.
arXiv Detail & Related papers (2024-05-26T15:17:28Z) - Clothes-Changing Person Re-Identification with Feasibility-Aware Intermediary Matching [86.04494755636613]
Current clothes-changing person re-identification (re-id) approaches usually perform retrieval based on clothes-irrelevant features.<n>We propose a Feasibility-Aware Intermediary Matching (FAIM) framework to additionally utilize clothes-relevant features for retrieval.<n>Our method outperforms state-of-the-art methods on several widely-used clothes-changing re-id benchmarks.
arXiv Detail & Related papers (2024-04-15T06:58:09Z) - VILA: Learning Image Aesthetics from User Comments with Vision-Language
Pretraining [53.470662123170555]
We propose learning image aesthetics from user comments, and exploring vision-language pretraining methods to learn multimodal aesthetic representations.
Specifically, we pretrain an image-text encoder-decoder model with image-comment pairs, using contrastive and generative objectives to learn rich and generic aesthetic semantics without human labels.
Our results show that our pretrained aesthetic vision-language model outperforms prior works on image aesthetic captioning over the AVA-Captions dataset.
arXiv Detail & Related papers (2023-03-24T23:57:28Z) - Attribute-aware Explainable Complementary Clothing Recommendation [37.30129304097086]
This work aims to tackle the explainability challenge in fashion recommendation tasks by proposing a novel Attribute-aware Fashion Recommender (AFRec)
AFRec recommender assesses the outfit compatibility by explicitly leveraging the extracted attribute-level representations from each item's visual feature.
The attributes serve as the bridge between two fashion items, where we quantify the affinity of a pair of items through the learned compatibility between their attributes.
arXiv Detail & Related papers (2021-07-04T14:56:07Z) - Learning Diverse Fashion Collocation by Neural Graph Filtering [78.9188246136867]
We propose a novel fashion collocation framework, Neural Graph Filtering, that models a flexible set of fashion items via a graph neural network.
By applying symmetric operations on the edge vectors, this framework allows varying numbers of inputs/outputs and is invariant to their ordering.
We evaluate the proposed approach on three popular benchmarks, the Polyvore dataset, the Polyvore-D dataset, and our reorganized Amazon Fashion dataset.
arXiv Detail & Related papers (2020-03-11T16:17:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.