Multimodal Material Classification for Robots using Spectroscopy and
High Resolution Texture Imaging
- URL: http://arxiv.org/abs/2004.01160v2
- Date: Thu, 30 Jul 2020 19:45:38 GMT
- Title: Multimodal Material Classification for Robots using Spectroscopy and
High Resolution Texture Imaging
- Authors: Zackory Erickson, Eliot Xing, Bharat Srirangam, Sonia Chernova, and
Charles C. Kemp
- Abstract summary: We present a multimodal sensing technique, leveraging near-infrared spectroscopy and close-range high resolution texture imaging.
We show that this representation enables a robot to recognize materials with greater performance as compared to prior state-of-the-art approaches.
- Score: 14.458436940557924
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Material recognition can help inform robots about how to properly interact
with and manipulate real-world objects. In this paper, we present a multimodal
sensing technique, leveraging near-infrared spectroscopy and close-range high
resolution texture imaging, that enables robots to estimate the materials of
household objects. We release a dataset of high resolution texture images and
spectral measurements collected from a mobile manipulator that interacted with
144 household objects. We then present a neural network architecture that
learns a compact multimodal representation of spectral measurements and texture
images. When generalizing material classification to new objects, we show that
this multimodal representation enables a robot to recognize materials with
greater performance as compared to prior state-of-the-art approaches. Finally,
we present how a robot can combine this high resolution local sensing with
images from the robot's head-mounted camera to achieve accurate material
classification over a scene of objects on a table.
Related papers
- Teaching Unknown Objects by Leveraging Human Gaze and Augmented Reality
in Human-Robot Interaction [3.1473798197405953]
This dissertation aims to teach a robot unknown objects in the context of Human-Robot Interaction (HRI)
The combination of eye tracking and Augmented Reality created a powerful synergy that empowered the human teacher to communicate with the robot.
The robot's object detection capabilities exhibited comparable performance to state-of-the-art object detectors trained on extensive datasets.
arXiv Detail & Related papers (2023-12-12T11:34:43Z) - GM-NeRF: Learning Generalizable Model-based Neural Radiance Fields from
Multi-view Images [79.39247661907397]
We introduce an effective framework Generalizable Model-based Neural Radiance Fields to synthesize free-viewpoint images.
Specifically, we propose a geometry-guided attention mechanism to register the appearance code from multi-view 2D images to a geometry proxy.
arXiv Detail & Related papers (2023-03-24T03:32:02Z) - Neural Scene Representation for Locomotion on Structured Terrain [56.48607865960868]
We propose a learning-based method to reconstruct the local terrain for a mobile robot traversing urban environments.
Using a stream of depth measurements from the onboard cameras and the robot's trajectory, the estimates the topography in the robot's vicinity.
We propose a 3D reconstruction model that faithfully reconstructs the scene, despite the noisy measurements and large amounts of missing data coming from the blind spots of the camera arrangement.
arXiv Detail & Related papers (2022-06-16T10:45:17Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Robotic Testbed for Rendezvous and Optical Navigation: Multi-Source
Calibration and Machine Learning Use Cases [0.0]
This work presents the most recent advances of the Robotic Testbed for Rendezvous and Optical Navigation (TRON) at Stanford University.
The TRON facility consists of two 6 degrees-of-freedom KUKA robot arms and a set of Vicon motion track cameras to reconfigure an arbitrary relative pose between a camera and a target mockup model.
A comparative analysis of the synthetic and TRON simulated imageries is performed using a Convolutional Neural Network (CNN) pre-trained on the synthetic images.
arXiv Detail & Related papers (2021-08-12T04:38:50Z) - Visionary: Vision architecture discovery for robot learning [58.67846907923373]
We propose a vision-based architecture search algorithm for robot manipulation learning, which discovers interactions between low dimension action inputs and high dimensional visual inputs.
Our approach automatically designs architectures while training on the task - discovering novel ways of combining and attending image feature representations with actions as well as features from previous layers.
arXiv Detail & Related papers (2021-03-26T17:51:43Z) - Nothing But Geometric Constraints: A Model-Free Method for Articulated
Object Pose Estimation [89.82169646672872]
We propose an unsupervised vision-based system to estimate the joint configurations of the robot arm from a sequence of RGB or RGB-D images without knowing the model a priori.
We combine a classical geometric formulation with deep learning and extend the use of epipolar multi-rigid-body constraints to solve this task.
arXiv Detail & Related papers (2020-11-30T20:46:48Z) - A Multisensory Learning Architecture for Rotation-invariant Object
Recognition [0.0]
This study presents a multisensory machine learning architecture for object recognition by employing a novel dataset that was constructed with the iCub robot.
The proposed architecture combines convolutional neural networks to form representations (i.e., features) ford grayscale color images and a multi-layer perceptron algorithm to process depth data.
arXiv Detail & Related papers (2020-09-14T09:39:48Z) - The iCub multisensor datasets for robot and computer vision applications [0.7340017786387767]
This document presents novel datasets constructed by employing the iCub robot equipped with an additional depth sensor and color camera.
We used the robot to acquire color and depth information for 210 objects in different acquisition scenarios.
arXiv Detail & Related papers (2020-03-04T10:59:29Z) - Learning Depth With Very Sparse Supervision [57.911425589947314]
This paper explores the idea that perception gets coupled to 3D properties of the world via interaction with the environment.
We train a specialized global-local network architecture with what would be available to a robot interacting with the environment.
Experiments on several datasets show that, when ground truth is available even for just one of the image pixels, the proposed network can learn monocular dense depth estimation up to 22.5% more accurately than state-of-the-art approaches.
arXiv Detail & Related papers (2020-03-02T10:44:13Z) - Real-Time Object Detection and Recognition on Low-Compute Humanoid
Robots using Deep Learning [0.12599533416395764]
We describe a novel architecture that enables multiple low-compute NAO robots to perform real-time detection, recognition and localization of objects in its camera view.
The proposed algorithm for object detection and localization is an empirical modification of YOLOv3, based on indoor experiments in multiple scenarios.
The architecture also comprises of an effective end-to-end pipeline to feed the real-time frames from the camera feed to the neural net and use its results for guiding the robot.
arXiv Detail & Related papers (2020-01-20T05:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.