Few-Shot Object Detection and Viewpoint Estimation for Objects in the
Wild
- URL: http://arxiv.org/abs/2007.12107v2
- Date: Wed, 12 Oct 2022 14:46:41 GMT
- Title: Few-Shot Object Detection and Viewpoint Estimation for Objects in the
Wild
- Authors: Yang Xiao, Vincent Lepetit, Renaud Marlet
- Abstract summary: We tackle the problems of few-shot object detection and few-shot viewpoint estimation.
We demonstrate on both tasks the benefits of guiding the network prediction with class-representative features extracted from data.
Our method outperforms state-of-the-art methods by a large margin on a range of datasets.
- Score: 40.132988301147776
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Detecting objects and estimating their viewpoints in images are key tasks of
3D scene understanding. Recent approaches have achieved excellent results on
very large benchmarks for object detection and viewpoint estimation. However,
performances are still lagging behind for novel object categories with few
samples. In this paper, we tackle the problems of few-shot object detection and
few-shot viewpoint estimation. We demonstrate on both tasks the benefits of
guiding the network prediction with class-representative features extracted
from data in different modalities: image patches for object detection, and
aligned 3D models for viewpoint estimation. Despite its simplicity, our method
outperforms state-of-the-art methods by a large margin on a range of datasets,
including PASCAL and COCO for few-shot object detection, and Pascal3D+ and
ObjectNet3D for few-shot viewpoint estimation. Furthermore, when the 3D model
is not available, we introduce a simple category-agnostic viewpoint estimation
method by exploiting geometrical similarities and consistent pose labelling
across different classes. While it moderately reduces performance, this
approach still obtains better results than previous methods in this setting.
Last, for the first time, we tackle the combination of both few-shot tasks, on
three challenging benchmarks for viewpoint estimation in the wild, ObjectNet3D,
Pascal3D+ and Pix3D, showing very promising results.
Related papers
- LocaliseBot: Multi-view 3D object localisation with differentiable
rendering for robot grasping [9.690844449175948]
We focus on object pose estimation.
Our approach relies on three pieces of information: multiple views of the object, the camera's parameters at those viewpoints, and 3D CAD models of objects.
We show that the estimated object pose results in 99.65% grasp accuracy with the ground truth grasp candidates.
arXiv Detail & Related papers (2023-11-14T14:27:53Z) - PoseMatcher: One-shot 6D Object Pose Estimation by Deep Feature Matching [51.142988196855484]
We propose PoseMatcher, an accurate model free one-shot object pose estimator.
We create a new training pipeline for object to image matching based on a three-view system.
To enable PoseMatcher to attend to distinct input modalities, an image and a pointcloud, we introduce IO-Layer.
arXiv Detail & Related papers (2023-04-03T21:14:59Z) - NOPE: Novel Object Pose Estimation from a Single Image [67.11073133072527]
We propose an approach that takes a single image of a new object as input and predicts the relative pose of this object in new images without prior knowledge of the object's 3D model.
We achieve this by training a model to directly predict discriminative embeddings for viewpoints surrounding the object.
This prediction is done using a simple U-Net architecture with attention and conditioned on the desired pose, which yields extremely fast inference.
arXiv Detail & Related papers (2023-03-23T18:55:43Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - Analysis of voxel-based 3D object detection methods efficiency for
real-time embedded systems [93.73198973454944]
Two popular voxel-based 3D object detection methods are studied in this paper.
Our experiments show that these methods mostly fail to detect distant small objects due to the sparsity of the input point clouds at large distances.
Our findings suggest that a considerable part of the computations of existing methods is focused on locations of the scene that do not contribute with successful detection.
arXiv Detail & Related papers (2021-05-21T12:40:59Z) - Objects are Different: Flexible Monocular 3D Object Detection [87.82253067302561]
We propose a flexible framework for monocular 3D object detection which explicitly decouples the truncated objects and adaptively combines multiple approaches for object depth estimation.
Experiments demonstrate that our method outperforms the state-of-the-art method by relatively 27% for the moderate level and 30% for the hard level in the test set of KITTI benchmark.
arXiv Detail & Related papers (2021-04-06T07:01:28Z) - Few-shot Object Detection on Remote Sensing Images [11.40135025181393]
We introduce a few-shot learning-based method for object detection on remote sensing images.
We build our few-shot object detection model upon YOLOv3 architecture and develop a multi-scale object detection framework.
arXiv Detail & Related papers (2020-06-14T07:18:10Z) - Object as Hotspots: An Anchor-Free 3D Object Detection Approach via
Firing of Hotspots [37.16690737208046]
We argue for an approach opposite to existing methods using object-level anchors.
Inspired by compositional models, we propose an object as composition of its interior non-empty voxels, termed hotspots.
Based on OHS, we propose an anchor-free detection head with a novel ground truth assignment strategy.
arXiv Detail & Related papers (2019-12-30T03:02:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.