Orient Anything
- URL: http://arxiv.org/abs/2410.02101v2
- Date: Fri, 31 Jan 2025 21:50:54 GMT
- Title: Orient Anything
- Authors: Christopher Scarvelis, David Benhaim, Paul Zhang,
- Abstract summary: We introduce a two-stage orientation pipeline that achieves state of the art performance on up-axis estimation.
Unlike previous work, we train and evaluate our method on all of Shapenet rather than a subset of classes.
- Score: 4.342241136871849
- License:
- Abstract: Orientation estimation is a fundamental task in 3D shape analysis which consists of estimating a shape's orientation axes: its side-, up-, and front-axes. Using this data, one can rotate a shape into canonical orientation, where its orientation axes are aligned with the coordinate axes. Developing an orientation algorithm that reliably estimates complete orientations of general shapes remains an open problem. We introduce a two-stage orientation pipeline that achieves state of the art performance on up-axis estimation and further demonstrate its efficacy on full-orientation estimation, where one seeks all three orientation axes. Unlike previous work, we train and evaluate our method on all of Shapenet rather than a subset of classes. We motivate our engineering contributions by theory describing fundamental obstacles to orientation estimation for rotationally-symmetric shapes, and show how our method avoids these obstacles.
Related papers
- Rotation-Adaptive Point Cloud Domain Generalization via Intricate Orientation Learning [34.424450834358204]
We propose an innovative rotation-adaptive domain generalization framework for 3D point cloud analysis.
Our approach aims to alleviate orientational shifts by leveraging intricate samples in an iterative learning process.
We employ an orientation-aware contrastive learning framework that incorporates an orientation consistency loss and a margin separation loss.
arXiv Detail & Related papers (2025-02-04T11:46:32Z) - Orient Anything: Learning Robust Object Orientation Estimation from Rendering 3D Models [79.96917782423219]
Orient Anything is the first expert and foundational model designed to estimate object orientation in a single image.
By developing a pipeline to annotate the front face of 3D objects, we collect 2M images with precise orientation annotations.
Our model achieves state-of-the-art orientation estimation accuracy in both rendered and real images.
arXiv Detail & Related papers (2024-12-24T18:58:43Z) - ESCAPE: Equivariant Shape Completion via Anchor Point Encoding [79.59829525431238]
We introduce ESCAPE, a framework designed to achieve rotation-equivariant shape completion.
ESCAPE employs a distinctive encoding strategy by selecting anchor points from a shape and representing all points as a distance to all anchor points.
ESCAPE achieves robust, high-quality reconstructions across arbitrary rotations and translations.
arXiv Detail & Related papers (2024-12-01T20:05:14Z) - Category-Level 6D Object Pose Estimation with Flexible Vector-Based
Rotation Representation [51.67545893892129]
We propose a novel 3D graph convolution based pipeline for category-level 6D pose and size estimation from monocular RGB-D images.
We first design an orientation-aware autoencoder with 3D graph convolution for latent feature learning.
Then, to efficiently decode the rotation information from the latent feature, we design a novel flexible vector-based decomposable rotation representation.
arXiv Detail & Related papers (2022-12-09T02:13:43Z) - Deep Projective Rotation Estimation through Relative Supervision [31.05330535795121]
Deep learning has offered a way to develop image-based orientation estimators.
These estimators often require training on a large labeled dataset.
We propose a new algorithm for selfsupervised orientation estimation.
arXiv Detail & Related papers (2022-11-21T04:58:07Z) - Adjoint Rigid Transform Network: Task-conditioned Alignment of 3D Shapes [86.2129580231191]
Adjoint Rigid Transform (ART) Network is a neural module which can be integrated with a variety of 3D networks.
ART learns to rotate input shapes to a learned canonical orientation, which is crucial for a lot of tasks.
We will release our code and pre-trained models for further research.
arXiv Detail & Related papers (2021-02-01T20:58:45Z) - Amplifying the Anterior-Posterior Difference via Data Enhancement -- A
More Robust Deep Monocular Orientation Estimation Solution [7.540176446791261]
Existing deep-learning based monocular orientation estimation algorithms face the problem of confusion between the anterior and posterior parts of the objects.
We propose a pretraining method which focuses on predicting the left/right semicircle in which the orientation of the object is located.
Experiment results show that the proposed semicircle prediction enhances the accuracy of orientation estimation, and mitigates the problem stated above.
arXiv Detail & Related papers (2020-12-21T15:36:13Z) - Learning to Orient Surfaces by Self-supervised Spherical CNNs [15.554429755106332]
Defining and reliably finding a canonical orientation for 3D surfaces is key to many Computer Vision and Robotics applications.
We show the feasibility of learning a robust canonical orientation for surfaces represented as point clouds.
Our method learns such feature maps from raw data by a self-supervised training procedure and robustly selects a rotation to transform the input point cloud into a learned canonical orientation.
arXiv Detail & Related papers (2020-11-06T11:43:57Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z) - From Planes to Corners: Multi-Purpose Primitive Detection in Unorganized
3D Point Clouds [59.98665358527686]
We propose a new method for segmentation-free joint estimation of orthogonal planes.
Such unified scene exploration allows for multitudes of applications such as semantic plane detection or local and global scan alignment.
Our experiments demonstrate the validity of our approach in numerous scenarios from wall detection to 6D tracking.
arXiv Detail & Related papers (2020-01-21T06:51:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.