Predicting and Attending to Damaging Collisions for Placing Everyday
Objects in Photo-Realistic Simulations
- URL: http://arxiv.org/abs/2102.06507v1
- Date: Fri, 12 Feb 2021 13:21:45 GMT
- Title: Predicting and Attending to Damaging Collisions for Placing Everyday
Objects in Photo-Realistic Simulations
- Authors: Aly Magassouba, Komei Sugiura, Angelica Nakayama, Tsubasa Hirakawa,
Takayoshi Yamashita, Hironobu Fujiyoshi, Hisashi Kawai
- Abstract summary: We show that a rule-based approach that uses plane detection, to detect free areas, performs poorly.
We develop PonNet, which has multimodal attention branches and a self-attention mechanism to predict damaging collisions.
Our method can visualize the risk of damaging collisions, which is convenient because it enables the user to understand the risk.
- Score: 27.312610846200187
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Placing objects is a fundamental task for domestic service robots (DSRs).
Thus, inferring the collision-risk before a placing motion is crucial for
achieving the requested task. This problem is particularly challenging because
it is necessary to predict what happens if an object is placed in a cluttered
designated area. We show that a rule-based approach that uses plane detection,
to detect free areas, performs poorly. To address this, we develop PonNet,
which has multimodal attention branches and a self-attention mechanism to
predict damaging collisions, based on RGBD images. Our method can visualize the
risk of damaging collisions, which is convenient because it enables the user to
understand the risk. For this purpose, we build and publish an original dataset
that contains 12,000 photo-realistic images of specific placing areas, with
daily life objects, in home environments. The experimental results show that
our approach improves accuracy compared with the baseline methods.
Related papers
- OOSTraj: Out-of-Sight Trajectory Prediction With Vision-Positioning Denoising [49.86409475232849]
Trajectory prediction is fundamental in computer vision and autonomous driving.
Existing approaches in this field often assume precise and complete observational data.
We present a novel method for out-of-sight trajectory prediction that leverages a vision-positioning technique.
arXiv Detail & Related papers (2024-04-02T18:30:29Z) - Unsupervised learning based object detection using Contrastive Learning [6.912349403119665]
We introduce a groundbreaking method for training single-stage object detectors through unsupervised/self-supervised learning.
Our state-of-the-art approach has the potential to revolutionize the labeling process, substantially reducing the time and cost associated with manual annotation.
We pioneer the concept of intra-image contrastive learning alongside inter-image counterparts, enabling the acquisition of crucial location information.
arXiv Detail & Related papers (2024-02-21T01:44:15Z) - STRIDE: Street View-based Environmental Feature Detection and Pedestrian
Collision Prediction [1.002773173311891]
We introduce a built environment detection task in large-scale panoramic images and a detection-based pedestrian collision frequency prediction task.
Our experiments demonstrate a significant correlation between object detection of built environment elements and pedestrian collision frequency prediction.
arXiv Detail & Related papers (2023-08-25T05:25:01Z) - You Only Crash Once: Improved Object Detection for Real-Time,
Sim-to-Real Hazardous Terrain Detection and Classification for Autonomous
Planetary Landings [7.201292864036088]
A cheap and effective way of detecting hazardous terrain is through the use of visual cameras.
Traditional techniques for visual hazardous terrain detection focus on template matching and registration to pre-built hazard maps.
We introduce You Only Crash Once (YOCO), a deep learning-based visual hazardous terrain detection and classification technique.
arXiv Detail & Related papers (2023-03-08T21:11:51Z) - To Make Yourself Invisible with Adversarial Semantic Contours [47.755808439588094]
Adversarial Semantic Contour (ASC) is an estimate of a Bayesian formulation of sparse attack with a deceived prior of object contour.
We show that ASC can corrupt the prediction of 9 modern detectors with different architectures.
We conclude with cautions about contour being the common weakness of object detectors with various architecture.
arXiv Detail & Related papers (2023-03-01T07:22:39Z) - Adversarially-Aware Robust Object Detector [85.10894272034135]
We propose a Robust Detector (RobustDet) based on adversarially-aware convolution to disentangle gradients for model learning on clean and adversarial images.
Our model effectively disentangles gradients and significantly enhances the detection robustness with maintaining the detection ability on clean images.
arXiv Detail & Related papers (2022-07-13T13:59:59Z) - Self-supervised Human Detection and Segmentation via Multi-view
Consensus [116.92405645348185]
We propose a multi-camera framework in which geometric constraints are embedded in the form of multi-view consistency during training.
We show that our approach outperforms state-of-the-art self-supervised person detection and segmentation techniques on images that visually depart from those of standard benchmarks.
arXiv Detail & Related papers (2020-12-09T15:47:21Z) - Self-supervised Segmentation via Background Inpainting [96.10971980098196]
We introduce a self-supervised detection and segmentation approach that can work with single images captured by a potentially moving camera.
We exploit a self-supervised loss function that we exploit to train a proposal-based segmentation network.
We apply our method to human detection and segmentation in images that visually depart from those of standard benchmarks and outperform existing self-supervised methods.
arXiv Detail & Related papers (2020-11-11T08:34:40Z) - Risk-Averse MPC via Visual-Inertial Input and Recurrent Networks for
Online Collision Avoidance [95.86944752753564]
We propose an online path planning architecture that extends the model predictive control (MPC) formulation to consider future location uncertainties.
Our algorithm combines an object detection pipeline with a recurrent neural network (RNN) which infers the covariance of state estimates.
The robustness of our methods is validated on complex quadruped robot dynamics and can be generally applied to most robotic platforms.
arXiv Detail & Related papers (2020-07-28T07:34:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.