Data Efficient Training with Imbalanced Label Sample Distribution for
Fashion Detection
- URL: http://arxiv.org/abs/2305.04379v5
- Date: Tue, 6 Jun 2023 07:33:13 GMT
- Title: Data Efficient Training with Imbalanced Label Sample Distribution for
Fashion Detection
- Authors: Xin Shen, Praful Agrawal, Zhongwei Cheng
- Abstract summary: We propose a state-of-the-art weighted objective function to boost the performance of deep neural networks (DNNs) for multi-label classification with long-tailed data distribution.
Our experiments involve image-based attribute classification of fashion apparels, and the results demonstrate favorable performance for the new weighting method.
- Score: 5.912870746288055
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Multi-label classification models have a wide range of applications in
E-commerce, including visual-based label predictions and language-based
sentiment classifications. A major challenge in achieving satisfactory
performance for these tasks in the real world is the notable imbalance in data
distribution. For instance, in fashion attribute detection, there may be only
six 'puff sleeve' clothes among 1000 products in most E-commerce fashion
catalogs. To address this issue, we explore more data-efficient model training
techniques rather than acquiring a huge amount of annotations to collect
sufficient samples, which is neither economic nor scalable. In this paper, we
propose a state-of-the-art weighted objective function to boost the performance
of deep neural networks (DNNs) for multi-label classification with long-tailed
data distribution. Our experiments involve image-based attribute classification
of fashion apparels, and the results demonstrate favorable performance for the
new weighting method compared to non-weighted and inverse-frequency-based
weighting mechanisms. We further evaluate the robustness of the new weighting
mechanism using two popular fashion attribute types in today's fashion
industry: sleevetype and archetype.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - Balancing Efficiency vs. Effectiveness and Providing Missing Label
Robustness in Multi-Label Stream Classification [3.97048491084787]
We propose a neural network-based approach to high-dimensional multi-label classification.
Our model uses a selective concept drift adaptation mechanism that makes it suitable for a non-stationary environment.
We adapt our model to an environment with missing labels using a simple yet effective imputation strategy.
arXiv Detail & Related papers (2023-10-01T13:23:37Z) - Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot
Text Classification Tasks [75.42002070547267]
We propose a self evolution learning (SE) based mixup approach for data augmentation in text classification.
We introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up.
arXiv Detail & Related papers (2023-05-22T23:43:23Z) - Consistent Text Categorization using Data Augmentation in e-Commerce [1.558017967663767]
We propose a new framework for consistent text categorization.
Our goal is to improve the model's consistency while maintaining its production-level performance.
arXiv Detail & Related papers (2023-05-09T12:47:28Z) - DETR-based Layered Clothing Segmentation and Fine-Grained Attribute
Recognition [10.924683447616273]
A new DEtection TRansformer (DETR) based method is proposed to segment and recognize fine-grained attributes of ensemble clothing instances with high accuracy.
We train our model on the Fashionpedia dataset and demonstrate our method surpasses SOTA models in tasks of layered clothing segmentation and fine-grained attribute recognition.
arXiv Detail & Related papers (2023-04-17T09:34:48Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Comparing Test Sets with Item Response Theory [53.755064720563]
We evaluate 29 datasets using predictions from 18 pretrained Transformer models on individual test examples.
We find that Quoref, HellaSwag, and MC-TACO are best suited for distinguishing among state-of-the-art models.
We also observe span selection task format, which is used for QA datasets like QAMR or SQuAD2.0, is effective in differentiating between strong and weak models.
arXiv Detail & Related papers (2021-06-01T22:33:53Z) - Exploiting Transductive Property of Graph Convolutional Neural Networks
with Less Labeling Effort [0.0]
The developing GCN model has made significant experimental contributions with Convolution filters applied to graph data.
Due to its transductive property, all of the data samples, which is partially labeled, are given as input to the model.
arXiv Detail & Related papers (2021-05-01T05:33:31Z) - Overcoming Classifier Imbalance for Long-tail Object Detection with
Balanced Group Softmax [88.11979569564427]
We provide the first systematic analysis on the underperformance of state-of-the-art models in front of long-tail distribution.
We propose a novel balanced group softmax (BAGS) module for balancing the classifiers within the detection frameworks through group-wise training.
Extensive experiments on the very recent long-tail large vocabulary object recognition benchmark LVIS show that our proposed BAGS significantly improves the performance of detectors.
arXiv Detail & Related papers (2020-06-18T10:24:26Z) - Learning Diverse Fashion Collocation by Neural Graph Filtering [78.9188246136867]
We propose a novel fashion collocation framework, Neural Graph Filtering, that models a flexible set of fashion items via a graph neural network.
By applying symmetric operations on the edge vectors, this framework allows varying numbers of inputs/outputs and is invariant to their ordering.
We evaluate the proposed approach on three popular benchmarks, the Polyvore dataset, the Polyvore-D dataset, and our reorganized Amazon Fashion dataset.
arXiv Detail & Related papers (2020-03-11T16:17:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.