Category-level Shape Estimation for Densely Cluttered Objects
- URL: http://arxiv.org/abs/2302.11983v1
- Date: Thu, 23 Feb 2023 13:00:17 GMT
- Title: Category-level Shape Estimation for Densely Cluttered Objects
- Authors: Zhenyu Wu, Ziwei Wang, Jiwen Lu and Haibin Yan
- Abstract summary: We propose a category-level shape estimation method for densely cluttered objects.
Our framework partitions each object in the clutter via the multi-view visual information fusion.
Experiments in the simulated environment and real world show that our method achieves high shape estimation accuracy.
- Score: 94.64287790278887
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurately estimating the shape of objects in dense clutters makes important
contribution to robotic packing, because the optimal object arrangement
requires the robot planner to acquire shape information of all existed objects.
However, the objects for packing are usually piled in dense clutters with
severe occlusion, and the object shape varies significantly across different
instances for the same category. They respectively cause large object
segmentation errors and inaccurate shape recovery on unseen instances, which
both degrade the performance of shape estimation during deployment. In this
paper, we propose a category-level shape estimation method for densely
cluttered objects. Our framework partitions each object in the clutter via the
multi-view visual information fusion to achieve high segmentation accuracy, and
the instance shape is recovered by deforming the category templates with
diverse geometric transformations to obtain strengthened generalization
ability. Specifically, we first collect the multi-view RGB-D images of the
object clutters for point cloud reconstruction. Then we fuse the feature maps
representing the visual information of multi-view RGB images and the pixel
affinity learned from the clutter point cloud, where the acquired instance
segmentation masks of multi-view RGB images are projected to partition the
clutter point cloud. Finally, the instance geometry information is obtained
from the partially observed instance point cloud and the corresponding category
template, and the deformation parameters regarding the template are predicted
for shape estimation. Experiments in the simulated environment and real world
show that our method achieves high shape estimation accuracy for densely
cluttered everyday objects with various shapes.
Related papers
- KP-RED: Exploiting Semantic Keypoints for Joint 3D Shape Retrieval and Deformation [87.23575166061413]
KP-RED is a unified KeyPoint-driven REtrieval and Deformation framework.
It takes object scans as input and jointly retrieves and deforms the most geometrically similar CAD models.
arXiv Detail & Related papers (2024-03-15T08:44:56Z) - Generative Category-Level Shape and Pose Estimation with Semantic
Primitives [27.692997522812615]
We propose a novel framework for category-level object shape and pose estimation from a single RGB-D image.
To handle the intra-category variation, we adopt a semantic primitive representation that encodes diverse shapes into a unified latent space.
We show that the proposed method achieves SOTA pose estimation performance and better generalization in the real-world dataset.
arXiv Detail & Related papers (2022-10-03T17:51:54Z) - Learning to Complete Object Shapes for Object-level Mapping in Dynamic
Scenes [30.500198859451434]
We propose a novel object-level mapping system that can simultaneously segment, track, and reconstruct objects in dynamic scenes.
It can further predict and complete their full geometries by conditioning on reconstructions from depth inputs and a category-level shape prior.
We evaluate its effectiveness by quantitatively and qualitatively testing it in both synthetic and real-world sequences.
arXiv Detail & Related papers (2022-08-09T22:56:33Z) - ELLIPSDF: Joint Object Pose and Shape Optimization with a Bi-level
Ellipsoid and Signed Distance Function Description [9.734266860544663]
This paper proposes an expressive yet compact model for joint object pose and shape optimization.
It infers an object-level map from multi-view RGB-D camera observations.
Our approach is evaluated on the large-scale real-world ScanNet dataset and compared against state-of-the-art methods.
arXiv Detail & Related papers (2021-08-01T03:07:31Z) - Localization and Mapping using Instance-specific Mesh Models [12.235379548921061]
This paper focuses on building semantic maps, containing object poses and shapes, using a monocular camera.
Our contribution is an instance-specific mesh model of object shape that can be optimized online based on semantic information extracted from camera images.
arXiv Detail & Related papers (2021-03-08T00:24:23Z) - From Points to Multi-Object 3D Reconstruction [71.17445805257196]
We propose a method to detect and reconstruct multiple 3D objects from a single RGB image.
A keypoint detector localizes objects as center points and directly predicts all object properties, including 9-DoF bounding boxes and 3D shapes.
The presented approach performs lightweight reconstruction in a single-stage, it is real-time capable, fully differentiable and end-to-end trainable.
arXiv Detail & Related papers (2020-12-21T18:52:21Z) - 3D Object Classification on Partial Point Clouds: A Practical
Perspective [91.81377258830703]
A point cloud is a popular shape representation adopted in 3D object classification.
This paper introduces a practical setting to classify partial point clouds of object instances under any poses.
A novel algorithm in an alignment-classification manner is proposed in this paper.
arXiv Detail & Related papers (2020-12-18T04:00:56Z) - Continuous Surface Embeddings [76.86259029442624]
We focus on the task of learning and representing dense correspondences in deformable object categories.
We propose a new, learnable image-based representation of dense correspondences.
We demonstrate that the proposed approach performs on par or better than the state-of-the-art methods for dense pose estimation for humans.
arXiv Detail & Related papers (2020-11-24T22:52:15Z) - Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation [62.618227434286]
We present a novel learning approach to recover the 6D poses and sizes of unseen object instances from an RGB-D image.
We propose a deep network to reconstruct the 3D object model by explicitly modeling the deformation from a pre-learned categorical shape prior.
arXiv Detail & Related papers (2020-07-16T16:45:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.