Towards Scale Balanced 6-DoF Grasp Detection in Cluttered Scenes
- URL: http://arxiv.org/abs/2212.05275v1
- Date: Sat, 10 Dec 2022 11:31:12 GMT
- Title: Towards Scale Balanced 6-DoF Grasp Detection in Cluttered Scenes
- Authors: Haoxiang Ma and Di Huang
- Abstract summary: We propose a novel approach to especially address the difficulty in dealing with small-scale samples.
A Multi-scale Cylinder Grouping (MsCG) module is presented to enhance local geometry representation.
Noisy-clean Mix (NcM) data augmentation is introduced to facilitate training.
- Score: 19.25678039613183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we focus on the problem of feature learning in the presence of
scale imbalance for 6-DoF grasp detection and propose a novel approach to
especially address the difficulty in dealing with small-scale samples. A
Multi-scale Cylinder Grouping (MsCG) module is presented to enhance local
geometry representation by combining multi-scale cylinder features and global
context. Moreover, a Scale Balanced Learning (SBL) loss and an Object Balanced
Sampling (OBS) strategy are designed, where SBL enlarges the gradients of the
samples whose scales are in low frequency by apriori weights while OBS captures
more points on small-scale objects with the help of an auxiliary segmentation
network. They alleviate the influence of the uneven distribution of grasp
scales in training and inference respectively. In addition, Noisy-clean Mix
(NcM) data augmentation is introduced to facilitate training, aiming to bridge
the domain gap between synthetic and raw scenes in an efficient way by
generating more data which mix them into single ones at instance-level.
Extensive experiments are conducted on the GraspNet-1Billion benchmark and
competitive results are reached with significant gains on small-scale cases.
Besides, the performance of real-world grasping highlights its generalization
ability. Our code is available at
https://github.com/mahaoxiang822/Scale-Balanced-Grasp.
Related papers
- Unbiased Max-Min Embedding Classification for Transductive Few-Shot Learning: Clustering and Classification Are All You Need [83.10178754323955]
Few-shot learning enables models to generalize from only a few labeled examples.
We propose the Unbiased Max-Min Embedding Classification (UMMEC) Method, which addresses the key challenges in few-shot learning.
Our method significantly improves classification performance with minimal labeled data, advancing the state-of-the-art in annotatedL.
arXiv Detail & Related papers (2025-03-28T07:23:07Z) - Scale Efficient Training for Large Datasets [27.28640920242675]
To remove low-value samples, SeTa first performs random pruning to eliminate redundant samples, then clusters the remaining samples according to their learning difficulty measured by loss.
SeTa reduces training costs by up to 50% while maintaining or improving performance, with minimal degradation even at 70% cost reduction.
arXiv Detail & Related papers (2025-03-17T17:13:43Z) - AutoScale: Scale-Aware Data Mixing for Pre-Training LLMs [61.13296177652599]
We show that data mixtures that perform well at smaller scales may not retain their advantage at larger scales.
We propose AutoScale, a two-stage, scale-aware data composition framework.
arXiv Detail & Related papers (2024-07-29T17:06:30Z) - Multi-Scale Denoising in the Feature Space for Low-Light Instance Segmentation [2.642212767247493]
Instance segmentation for low-light imagery remains largely unexplored.
Our proposed method implements weighted non-local blocks (wNLB) in the feature extractor.
We introduce additional learnable weights at each layer in order to enhance the network's adaptability to real-world noise characteristics.
arXiv Detail & Related papers (2024-02-28T13:07:16Z) - Scale Equalization for Multi-Level Feature Fusion [8.541075075344438]
We find that multi-level features from parallel branches are on different scales.
The scale disequilibrium is a universal and unwanted flaw that leads to detrimental gradient descent.
We propose injecting scale equalizers to achieve scale equilibrium across multi-level features after bilinear upsampling.
arXiv Detail & Related papers (2024-02-02T05:25:51Z) - Combating Representation Learning Disparity with Geometric Harmonization [50.29859682439571]
We propose a novel Geometric Harmonization (GH) method to encourage category-level uniformity in representation learning.
Our proposal does not alter the setting of SSL and can be easily integrated into existing methods in a low-cost manner.
arXiv Detail & Related papers (2023-10-26T17:41:11Z) - Deep Boosting Multi-Modal Ensemble Face Recognition with Sample-Level
Weighting [11.39204323420108]
Deep convolutional neural networks have achieved remarkable success in face recognition.
The current training benchmarks exhibit an imbalanced quality distribution.
This poses issues for generalization on hard samples since they are underrepresented during training.
Inspired by the well-known AdaBoost, we propose a sample-level weighting approach to incorporate the importance of different samples into the FR loss.
arXiv Detail & Related papers (2023-08-18T01:44:54Z) - L-DAWA: Layer-wise Divergence Aware Weight Aggregation in Federated
Self-Supervised Visual Representation Learning [14.888569402903562]
Integration of self-supervised learning (SSL) and federated learning (FL) into one coherent system can potentially offer data privacy guarantees.
We propose a new aggregation strategy termed Layer-wise Divergence Aware Weight Aggregation (L-DAWA) to mitigate the influence of client bias and divergence during FL aggregation.
arXiv Detail & Related papers (2023-07-14T15:07:30Z) - Scale-Equivalent Distillation for Semi-Supervised Object Detection [57.59525453301374]
Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals.
We analyze the challenges these methods meet with the empirical experiment results.
We introduce a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance.
arXiv Detail & Related papers (2022-03-23T07:33:37Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Suppressing Mislabeled Data via Grouping and Self-Attention [60.14212694011875]
Deep networks achieve excellent results on large-scale clean data but degrade significantly when learning from noisy labels.
This paper proposes a conceptually simple yet efficient training block, termed as Attentive Feature Mixup (AFM)
It allows paying more attention to clean samples and less to mislabeled ones via sample interactions in small groups.
arXiv Detail & Related papers (2020-10-29T13:54:16Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z) - Crowd Counting via Hierarchical Scale Recalibration Network [61.09833400167511]
We propose a novel Hierarchical Scale Recalibration Network (HSRNet) to tackle the task of crowd counting.
HSRNet models rich contextual dependencies and recalibrating multiple scale-associated information.
Our approach can ignore various noises selectively and focus on appropriate crowd scales automatically.
arXiv Detail & Related papers (2020-03-07T10:06:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.