Online Self-Supervised Learning for Object Picking: Detecting Optimum
Grasping Position using a Metric Learning Approach
- URL: http://arxiv.org/abs/2003.03717v1
- Date: Sun, 8 Mar 2020 04:36:24 GMT
- Title: Online Self-Supervised Learning for Object Picking: Detecting Optimum
Grasping Position using a Metric Learning Approach
- Authors: Kanata Suzuki, Yasuto Yokota, Yuzi Kanazawa, Tomoyoshi Takebayashi
- Abstract summary: The optimal grasping position of an individual object is determined from the grasping score.
The proposed online self-supervised learning method employs two deep neural networks.
- Score: 0.757024681220677
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised learning methods are attractive candidates for automatic
object picking. However, the trial samples lack the complete ground truth
because the observable parts of the agent are limited. That is, the information
contained in the trial samples is often insufficient to learn the specific
grasping position of each object. Consequently, the training falls into a local
solution, and the grasp positions learned by the robot are independent of the
state of the object. In this study, the optimal grasping position of an
individual object is determined from the grasping score, defined as the
distance in the feature space obtained using metric learning. The closeness of
the solution to the pre-designed optimal grasping position was evaluated in
trials. The proposed method incorporates two types of feedback control: one
feedback enlarges the grasping score when the grasping position approaches the
optimum; the other reduces the negative feedback of the potential grasping
positions among the grasping candidates. The proposed online self-supervised
learning method employs two deep neural networks. : SSD that detects the
grasping position of an object, and Siamese networks (SNs) that evaluate the
trial sample using the similarity of two input data in the feature space. Our
method embeds the relation of each grasping position as feature vectors by
training the trial samples and a few pre-samples indicating the optimum
grasping position. By incorporating the grasping score based on the feature
space of SNs into the SSD training process, the method preferentially trains
the optimum grasping position. In the experiment, the proposed method achieved
a higher success rate than the baseline method using simple teaching signals.
And the grasping scores in the feature space of the SNs accurately represented
the grasping positions of the objects.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - Referee-Meta-Learning for Fast Adaptation of Locational Fairness [26.770426062329165]
We propose a locational meta-referee (Meta-Ref) to oversee the few-shot meta-training and meta-testing of a deep neural network.
We show that Meta-Ref can improve locational fairness while keeping the overall prediction quality at a similar level.
arXiv Detail & Related papers (2024-02-20T21:09:04Z) - Optimal Parameter and Neuron Pruning for Out-of-Distribution Detection [36.4610463573214]
We propose an textbfOptimal textbfParameter and textbfNeuron textbfPruning (textbfOPNP) approach to detect out-of-distribution (OOD) samples.
Our proposal is training-free, compatible with other post-hoc methods, and exploring the information from all training data.
arXiv Detail & Related papers (2024-02-04T07:31:06Z) - Class-Imbalanced Semi-Supervised Learning for Large-Scale Point Cloud
Semantic Segmentation via Decoupling Optimization [64.36097398869774]
Semi-supervised learning (SSL) has been an active research topic for large-scale 3D scene understanding.
The existing SSL-based methods suffer from severe training bias due to class imbalance and long-tail distributions of the point cloud data.
We introduce a new decoupling optimization framework, which disentangles feature representation learning and classifier in an alternative optimization manner to shift the bias decision boundary effectively.
arXiv Detail & Related papers (2024-01-13T04:16:40Z) - VL4Pose: Active Learning Through Out-Of-Distribution Detection For Pose
Estimation [79.50280069412847]
We introduce VL4Pose, a first principles approach for active learning through out-of-distribution detection.
Our solution involves modelling the pose through a simple parametric Bayesian network trained via maximum likelihood estimation.
We perform qualitative and quantitative experiments on three datasets: MPII, LSP and ICVL, spanning human and hand pose estimation.
arXiv Detail & Related papers (2022-10-12T09:03:55Z) - CheckSel: Efficient and Accurate Data-valuation Through Online
Checkpoint Selection [3.321404824316694]
We propose a novel 2-phase solution to the problem of data valuation and subset selection.
Phase 1 selects representative checkpoints from an SGD-like training algorithm, which are used in phase-2 to estimate the approximate training data values.
Experimental results show the proposed algorithm outperforms recent baseline methods by up to 30% in terms of test accuracy.
arXiv Detail & Related papers (2022-03-14T02:06:52Z) - Active Learning for Deep Visual Tracking [51.5063680734122]
Convolutional neural networks (CNNs) have been successfully applied to the single target tracking task in recent years.
In this paper, we propose an active learning method for deep visual tracking, which selects and annotates the unlabeled samples to train the deep CNNs model.
Under the guidance of active learning, the tracker based on the trained deep CNNs model can achieve competitive tracking performance while reducing the labeling cost.
arXiv Detail & Related papers (2021-10-17T11:47:56Z) - SIMPLE: SIngle-network with Mimicking and Point Learning for Bottom-up
Human Pose Estimation [81.03485688525133]
We propose a novel multi-person pose estimation framework, SIngle-network with Mimicking and Point Learning for Bottom-up Human Pose Estimation (SIMPLE)
Specifically, in the training process, we enable SIMPLE to mimic the pose knowledge from the high-performance top-down pipeline.
Besides, SIMPLE formulates human detection and pose estimation as a unified point learning framework to complement each other in single-network.
arXiv Detail & Related papers (2021-04-06T13:12:51Z) - Learning Neural Network Subspaces [74.44457651546728]
Recent observations have advanced our understanding of the neural network optimization landscape.
With a similar computational cost as training one model, we learn lines, curves, and simplexes of high-accuracy neural networks.
With a similar computational cost as training one model, we learn lines, curves, and simplexes of high-accuracy neural networks.
arXiv Detail & Related papers (2021-02-20T23:26:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.