Basket-based Softmax
- URL: http://arxiv.org/abs/2201.09308v1
- Date: Sun, 23 Jan 2022 16:43:29 GMT
- Title: Basket-based Softmax
- Authors: Qiang Meng, Xinqian Gu, Xiaqing Xu, Feng Zhou
- Abstract summary: We propose a novel mining-during-training strategy called Basket-based Softmax (BBS)
For each training sample, we simultaneously adopt similarity scores as the clue to mining negative classes from other datasets.
We demonstrate the efficiency and superiority of the BBS on the tasks of face recognition and re-identification, with both simulated and real-world datasets.
- Score: 12.744577044692276
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Softmax-based losses have achieved state-of-the-art performances on various
tasks such as face recognition and re-identification. However, these methods
highly relied on clean datasets with global labels, which limits their usage in
many real-world applications. An important reason is that merging and
organizing datasets from various temporal and spatial scenarios is usually not
realistic, as noisy labels can be introduced and exponential-increasing
resources are required. To address this issue, we propose a novel
mining-during-training strategy called Basket-based Softmax (BBS) as well as
its parallel version to effectively train models on multiple datasets in an
end-to-end fashion. Specifically, for each training sample, we simultaneously
adopt similarity scores as the clue to mining negative classes from other
datasets, and dynamically add them to assist the learning of discriminative
features. Experimentally, we demonstrate the efficiency and superiority of the
BBS on the tasks of face recognition and re-identification, with both simulated
and real-world datasets.
Related papers
- Continual Learning for Multimodal Data Fusion of a Soft Gripper [1.0589208420411014]
A model trained on one data modality often fails when tested with a different modality.
We introduce a continual learning algorithm capable of incrementally learning different data modalities.
We evaluate the algorithm's effectiveness on a challenging custom multimodal dataset.
arXiv Detail & Related papers (2024-09-20T09:53:27Z) - LocoMuJoCo: A Comprehensive Imitation Learning Benchmark for Locomotion [20.545058017790428]
Imitation Learning holds great promise for enabling agile locomotion in embodied agents.
We present a novel benchmark designed to facilitate rigorous evaluation and comparison of IL algorithms.
This benchmark encompasses a diverse set of environments, including quadrupeds, bipeds, and musculoskeletal human models.
arXiv Detail & Related papers (2023-11-04T19:41:50Z) - Exploring Data Redundancy in Real-world Image Classification through
Data Selection [20.389636181891515]
Deep learning models often require large amounts of data for training, leading to increased costs.
We present two data valuation metrics based on Synaptic Intelligence and gradient norms, respectively, to study redundancy in real-world image data.
Online and offline data selection algorithms are then proposed via clustering and grouping based on the examined data values.
arXiv Detail & Related papers (2023-06-25T03:31:05Z) - STREAMLINE: Streaming Active Learning for Realistic Multi-Distributional
Settings [2.580765958706854]
STREAMLINE is a novel streaming active learning framework that mitigates scenario-driven slice imbalance in working labeled data.
We evaluate STREAMLINE on real-world streaming scenarios for image classification and object detection tasks.
arXiv Detail & Related papers (2023-05-18T02:01:45Z) - Multi-Environment Pretraining Enables Transfer to Action Limited
Datasets [129.24823721649028]
In reinforcement learning, available data of decision making is often not annotated with actions.
We propose combining large but sparsely-annotated datasets from a emphtarget environment of interest with fully-annotated datasets from various other emphsource environments.
We show that utilizing even one additional environment dataset of sequential labelled data during IDM pretraining gives rise to substantial improvements in generating action labels for unannotated sequences.
arXiv Detail & Related papers (2022-11-23T22:48:22Z) - Representation Learning for the Automatic Indexing of Sound Effects
Libraries [79.68916470119743]
We show that a task-specific but dataset-independent representation can successfully address data issues such as class imbalance, inconsistent class labels, and insufficient dataset size.
Detailed experimental results show the impact of metric learning approaches and different cross-dataset training methods on representational effectiveness.
arXiv Detail & Related papers (2022-08-18T23:46:13Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - Learning with Neighbor Consistency for Noisy Labels [69.83857578836769]
We present a method for learning from noisy labels that leverages similarities between training examples in feature space.
We evaluate our method on datasets evaluating both synthetic (CIFAR-10, CIFAR-100) and realistic (mini-WebVision, Clothing1M, mini-ImageNet-Red) noise.
arXiv Detail & Related papers (2022-02-04T15:46:27Z) - DomainMix: Learning Generalizable Person Re-Identification Without Human
Annotations [89.78473564527688]
This paper shows how to use labeled synthetic dataset and unlabeled real-world dataset to train a universal model.
In this way, human annotations are no longer required, and it is scalable to large and diverse real-world datasets.
Experimental results show that the proposed annotation-free method is more or less comparable to the counterpart trained with full human annotations.
arXiv Detail & Related papers (2020-11-24T08:15:53Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.