RFTrans: Leveraging Refractive Flow of Transparent Objects for Surface
Normal Estimation and Manipulation
- URL: http://arxiv.org/abs/2311.12398v2
- Date: Thu, 8 Feb 2024 01:44:30 GMT
- Title: RFTrans: Leveraging Refractive Flow of Transparent Objects for Surface
Normal Estimation and Manipulation
- Authors: Tutian Tang, Jiyu Liu, Jieyi Zhang, Haoyuan Fu, Wenqiang Xu, Cewu Lu
- Abstract summary: This paper introduces RFTrans, an RGB-D-based method for surface normal estimation and manipulation of transparent objects.
It integrates the RFNet, which predicts refractive flow, object mask, and boundaries, followed by the F2Net, which estimates surface normal from the refractive flow.
A real-world robot grasping task witnesses an 83% success rate, proving that refractive flow can help enable direct sim-to-real transfer.
- Score: 50.10282876199739
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Transparent objects are widely used in our daily lives, making it important
to teach robots to interact with them. However, it's not easy because the
reflective and refractive effects can make depth cameras fail to give accurate
geometry measurements. To solve this problem, this paper introduces RFTrans, an
RGB-D-based method for surface normal estimation and manipulation of
transparent objects. By leveraging refractive flow as an intermediate
representation, the proposed method circumvents the drawbacks of directly
predicting the geometry (e.g. surface normal) from images and helps bridge the
sim-to-real gap. It integrates the RFNet, which predicts refractive flow,
object mask, and boundaries, followed by the F2Net, which estimates surface
normal from the refractive flow. To make manipulation possible, a global
optimization module will take in the predictions, refine the raw depth, and
construct the point cloud with normal. An off-the-shelf analytical grasp
planning algorithm is followed to generate the grasp poses. We build a
synthetic dataset with physically plausible ray-tracing rendering techniques to
train the networks. Results show that the proposed method trained on the
synthetic dataset can consistently outperform the baseline method in both
synthetic and real-world benchmarks by a large margin. Finally, a real-world
robot grasping task witnesses an 83% success rate, proving that refractive flow
can help enable direct sim-to-real transfer. The code, data, and supplementary
materials are available at https://rftrans.robotflow.ai.
Related papers
- Aero-Nef: Neural Fields for Rapid Aircraft Aerodynamics Simulations [1.1932047172700866]
This paper presents a methodology to learn surrogate models of steady state fluid dynamics simulations on meshed domains.
The proposed models can be applied directly to unstructured domains for different flow conditions.
Remarkably, the method can perform inference five order of magnitude faster than the high fidelity solver on the RANS transonic airfoil dataset.
arXiv Detail & Related papers (2024-07-29T11:48:44Z) - Flatten Anything: Unsupervised Neural Surface Parameterization [76.4422287292541]
We introduce the Flatten Anything Model (FAM), an unsupervised neural architecture to achieve global free-boundary surface parameterization.
Compared with previous methods, our FAM directly operates on discrete surface points without utilizing connectivity information.
Our FAM is fully-automated without the need for pre-cutting and can deal with highly-complex topologies.
arXiv Detail & Related papers (2024-05-23T14:39:52Z) - Closing the Visual Sim-to-Real Gap with Object-Composable NeRFs [59.12526668734703]
We introduce Composable Object Volume NeRF (COV-NeRF), an object-composable NeRF model that is the centerpiece of a real-to-sim pipeline.
COV-NeRF extracts objects from real images and composes them into new scenes, generating photorealistic renderings and many types of 2D and 3D supervision.
arXiv Detail & Related papers (2024-03-07T00:00:02Z) - NeuralGF: Unsupervised Point Normal Estimation by Learning Neural
Gradient Function [55.86697795177619]
Normal estimation for 3D point clouds is a fundamental task in 3D geometry processing.
We introduce a new paradigm for learning neural gradient functions, which encourages the neural network to fit the input point clouds.
Our excellent results on widely used benchmarks demonstrate that our method can learn more accurate normals for both unoriented and oriented normal estimation tasks.
arXiv Detail & Related papers (2023-11-01T09:25:29Z) - NeRRF: 3D Reconstruction and View Synthesis for Transparent and Specular
Objects with Neural Refractive-Reflective Fields [23.099784003061618]
We introduce the refractive-reflective field to Neural radiance fields (NeRF)
NeRF uses straight rays and fails to deal with complicated light path changes caused by refraction and reflection.
We propose a virtual cone supersampling technique to achieve efficient and effective anti-aliasing.
arXiv Detail & Related papers (2023-09-22T17:59:12Z) - TransTouch: Learning Transparent Objects Depth Sensing Through Sparse
Touches [23.87056600709768]
We propose a method to finetune a stereo network with sparse depth labels automatically collected using a probing system with tactile feedback.
We show that our method can significantly improve real-world depth sensing accuracy, especially for transparent objects.
arXiv Detail & Related papers (2023-09-18T01:55:17Z) - TransCG: A Large-Scale Real-World Dataset for Transparent Object Depth
Completion and Grasping [46.6058840385155]
We contribute a large-scale real-world dataset for transparent object depth completion.
Our dataset contains 57,715 RGB-D images from 130 different scenes.
We propose an end-to-end depth completion network, which takes the RGB image and the inaccurate depth map as inputs and outputs a refined depth map.
arXiv Detail & Related papers (2022-02-17T06:50:20Z) - IFOR: Iterative Flow Minimization for Robotic Object Rearrangement [92.97142696891727]
IFOR, Iterative Flow Minimization for Robotic Object Rearrangement, is an end-to-end method for the problem of object rearrangement for unknown objects.
We show that our method applies to cluttered scenes, and in the real world, while training only on synthetic data.
arXiv Detail & Related papers (2022-02-01T20:03:56Z) - RGB-D Local Implicit Function for Depth Completion of Transparent
Objects [43.238923881620494]
Majority of perception methods in robotics require depth information provided by RGB-D cameras.
Standard 3D sensors fail to capture depth of transparent objects due to refraction and absorption of light.
We present a novel framework that can complete missing depth given noisy RGB-D input.
arXiv Detail & Related papers (2021-04-01T17:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.