The Sum of Its Parts: Visual Part Segmentation for Inertial Parameter
Identification of Manipulated Objects
- URL: http://arxiv.org/abs/2302.06685v2
- Date: Thu, 6 Jul 2023 02:18:21 GMT
- Title: The Sum of Its Parts: Visual Part Segmentation for Inertial Parameter
Identification of Manipulated Objects
- Authors: Philippe Nadeau, Matthew Giamou, Jonathan Kelly
- Abstract summary: Traditional methods for estimating the full set of inertial parameters rely on motions that are necessarily fast and unsafe.
We develop an inertial parameter identification algorithm that requires slow or'stop-and-go' motions only hence is ideally tailored for use around humans.
We demonstrate our algorithm by performing an intricate 'hammer balancing act' autonomously and online with a low-cost collaborative robotic arm.
- Score: 8.798250996263237
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To operate safely and efficiently alongside human workers, collaborative
robots (cobots) require the ability to quickly understand the dynamics of
manipulated objects. However, traditional methods for estimating the full set
of inertial parameters rely on motions that are necessarily fast and unsafe (to
achieve a sufficient signal-to-noise ratio). In this work, we take an
alternative approach: by combining visual and force-torque measurements, we
develop an inertial parameter identification algorithm that requires slow or
'stop-and-go' motions only, and hence is ideally tailored for use around
humans. Our technique, called Homogeneous Part Segmentation (HPS), leverages
the observation that man-made objects are often composed of distinct,
homogeneous parts. We combine a surface-based point clustering method with a
volumetric shape segmentation algorithm to quickly produce a part-level
segmentation of a manipulated object; the segmented representation is then used
by HPS to accurately estimate the object's inertial parameters. To benchmark
our algorithm, we create and utilize a novel dataset consisting of realistic
meshes, segmented point clouds, and inertial parameters for 20 common workshop
tools. Finally, we demonstrate the real-world performance and accuracy of HPS
by performing an intricate 'hammer balancing act' autonomously and online with
a low-cost collaborative robotic arm. Our code and dataset are open source and
freely available.
Related papers
- The Art of Imitation: Learning Long-Horizon Manipulation Tasks from Few Demonstrations [13.747258771184372]
There are several open challenges to applying TP-GMMs in the wild.
We factorize the robot's end-effector velocity into its direction and magnitude.
We then segment and sequence skills from complex demonstration trajectories.
Our approach enables learning complex manipulation tasks from just five demonstrations.
arXiv Detail & Related papers (2024-07-18T12:01:09Z) - MARS: Multimodal Active Robotic Sensing for Articulated Characterization [6.69660410213287]
We introduce MARS, a novel framework for articulated object characterization.
It features a multi-modal fusion module utilizing multi-scale RGB features to enhance point cloud features.
Our method effectively generalizes to real-world articulated objects, enhancing robot interactions.
arXiv Detail & Related papers (2024-07-01T11:32:39Z) - RPMArt: Towards Robust Perception and Manipulation for Articulated Objects [56.73978941406907]
It is essential that robots can exhibit robust perception and manipulation for articulated objects in real-world robotic applications.
We propose a framework towards Robust Perception and Manipulation for Articulated Objects ( RPMArt)
RPMArt learns to estimate the articulation parameters and manipulate the articulation part from the noisy point cloud.
arXiv Detail & Related papers (2024-03-24T05:55:39Z) - RISeg: Robot Interactive Object Segmentation via Body Frame-Invariant
Features [6.358423536732677]
We introduce a novel approach to correct inaccurate segmentation by using robot interaction and a designed body frame-invariant feature.
We demonstrate the effectiveness of our proposed interactive perception pipeline in accurately segmenting cluttered scenes by achieving an average object segmentation accuracy rate of 80.7%.
arXiv Detail & Related papers (2024-03-04T05:03:24Z) - ICGNet: A Unified Approach for Instance-Centric Grasping [42.92991092305974]
We introduce an end-to-end architecture for object-centric grasping.
We show the effectiveness of the proposed method by extensively evaluating it against state-of-the-art methods on synthetic datasets.
arXiv Detail & Related papers (2024-01-18T12:41:41Z) - Appearance-based Refinement for Object-Centric Motion Segmentation [95.80420062679104]
We introduce an appearance-based refinement method that leverages temporal consistency in video streams to correct inaccurate flow-based proposals.
Our approach involves a simple selection mechanism that identifies accurate flow-predicted masks as exemplars.
Its performance is evaluated on multiple video segmentation benchmarks, including DAVIS, YouTubeVOS, SegTrackv2, and FBMS-59.
arXiv Detail & Related papers (2023-12-18T18:59:51Z) - Multi-body SE(3) Equivariance for Unsupervised Rigid Segmentation and
Motion Estimation [49.56131393810713]
We present an SE(3) equivariant architecture and a training strategy to tackle this task in an unsupervised manner.
Our method excels in both model performance and computational efficiency, with only 0.25M parameters and 0.92G FLOPs.
arXiv Detail & Related papers (2023-06-08T22:55:32Z) - H-SAUR: Hypothesize, Simulate, Act, Update, and Repeat for Understanding
Object Articulations from Interactions [62.510951695174604]
"Hypothesize, Simulate, Act, Update, and Repeat" (H-SAUR) is a probabilistic generative framework that generates hypotheses about how objects articulate given input observations.
We show that the proposed model significantly outperforms the current state-of-the-art articulated object manipulation framework.
We further improve the test-time efficiency of H-SAUR by integrating a learned prior from learning-based vision models.
arXiv Detail & Related papers (2022-10-22T18:39:33Z) - Efficient Global-Local Memory for Real-time Instrument Segmentation of
Robotic Surgical Video [53.14186293442669]
We identify two important clues for surgical instrument perception, including local temporal dependency from adjacent frames and global semantic correlation in long-range duration.
We propose a novel dual-memory network (DMNet) to relate both global and local-temporal knowledge.
Our method largely outperforms the state-of-the-art works on segmentation accuracy while maintaining a real-time speed.
arXiv Detail & Related papers (2021-09-28T10:10:14Z) - Nothing But Geometric Constraints: A Model-Free Method for Articulated
Object Pose Estimation [89.82169646672872]
We propose an unsupervised vision-based system to estimate the joint configurations of the robot arm from a sequence of RGB or RGB-D images without knowing the model a priori.
We combine a classical geometric formulation with deep learning and extend the use of epipolar multi-rigid-body constraints to solve this task.
arXiv Detail & Related papers (2020-11-30T20:46:48Z) - "What's This?" -- Learning to Segment Unknown Objects from Manipulation
Sequences [27.915309216800125]
We present a novel framework for self-supervised grasped object segmentation with a robotic manipulator.
We propose a single, end-to-end trainable architecture which jointly incorporates motion cues and semantic knowledge.
Our method neither depends on any visual registration of a kinematic robot or 3D object models, nor on precise hand-eye calibration or any additional sensor data.
arXiv Detail & Related papers (2020-11-06T10:55:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.