ISAR: A Benchmark for Single- and Few-Shot Object Instance Segmentation
and Re-Identification
- URL: http://arxiv.org/abs/2311.02734v1
- Date: Sun, 5 Nov 2023 18:51:33 GMT
- Title: ISAR: A Benchmark for Single- and Few-Shot Object Instance Segmentation
and Re-Identification
- Authors: Nicolas Gorlo, Kenneth Blomqvist, Francesco Milano and Roland Siegwart
- Abstract summary: We propose ISAR, a benchmark and baseline method for single- and few-shot object identification.
We provide a semi-synthetic dataset of video sequences with ground-truth semantic annotations.
Our benchmark aligns with the emerging research trend of unifying Multi-Object Tracking, Video Object, and Re-identification.
- Score: 24.709695178222862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most object-level mapping systems in use today make use of an upstream
learned object instance segmentation model. If we want to teach them about a
new object or segmentation class, we need to build a large dataset and retrain
the system. To build spatial AI systems that can quickly be taught about new
objects, we need to effectively solve the problem of single-shot object
detection, instance segmentation and re-identification. So far there is neither
a method fulfilling all of these requirements in unison nor a benchmark that
could be used to test such a method. Addressing this, we propose ISAR, a
benchmark and baseline method for single- and few-shot object Instance
Segmentation And Re-identification, in an effort to accelerate the development
of algorithms that can robustly detect, segment, and re-identify objects from a
single or a few sparse training examples. We provide a semi-synthetic dataset
of video sequences with ground-truth semantic annotations, a standardized
evaluation pipeline, and a baseline method. Our benchmark aligns with the
emerging research trend of unifying Multi-Object Tracking, Video Object
Segmentation, and Re-identification.
Related papers
- Frustratingly Simple but Effective Zero-shot Detection and Segmentation:
Analysis and a Strong Baseline [45.03824571286718]
Methods for object detection and segmentation often require abundant instance-level annotations for training, which are time-consuming and expensive to collect.
To address this, the task of zero-shot object detection (or segmentation) aims at learning effective methods for identifying and localizing object instances for the categories that have no supervision available.
In this work, we extensively study these design choices, and carefully construct a simple yet extremely effective zero-shot recognition method.
arXiv Detail & Related papers (2023-02-14T20:00:30Z) - BURST: A Benchmark for Unifying Object Recognition, Segmentation and
Tracking in Video [58.71785546245467]
Multiple existing benchmarks involve tracking and segmenting objects in video.
There is little interaction between them due to the use of disparate benchmark datasets and metrics.
We propose BURST, a dataset which contains thousands of diverse videos with high-quality object masks.
All tasks are evaluated using the same data and comparable metrics, which enables researchers to consider them in unison.
arXiv Detail & Related papers (2022-09-25T01:27:35Z) - Sparse Instance Activation for Real-Time Instance Segmentation [72.23597664935684]
We propose a conceptually novel, efficient, and fully convolutional framework for real-time instance segmentation.
SparseInst has extremely fast inference speed and achieves 40 FPS and 37.9 AP on the COCO benchmark.
arXiv Detail & Related papers (2022-03-24T03:15:39Z) - Scalable Video Object Segmentation with Identification Mechanism [125.4229430216776]
This paper explores the challenges of achieving scalable and effective multi-object modeling for semi-supervised Video Object (VOS)
We present two innovative approaches, Associating Objects with Transformers (AOT) and Associating Objects with Scalable Transformers (AOST)
Our approaches surpass the state-of-the-art competitors and display exceptional efficiency and scalability consistently across all six benchmarks.
arXiv Detail & Related papers (2022-03-22T03:33:27Z) - Iterative Learning for Instance Segmentation [0.0]
State-of-the-art deep neural network models require large amounts of labeled data in order to perform well in this task.
We propose for the first time, an iterative learning and annotation method that is able to detect, segment and annotate instances in datasets composed of multiple similar objects.
Experiments on two different datasets show the validity of the approach in different applications related to visual inspection.
arXiv Detail & Related papers (2022-02-18T10:25:02Z) - Associating Objects with Transformers for Video Object Segmentation [74.51719591192787]
We propose an Associating Objects with Transformers (AOT) approach to match and decode multiple objects uniformly.
AOT employs an identification mechanism to associate multiple targets into the same high-dimensional embedding space.
We ranked 1st in the 3rd Large-scale Video Object Challenge.
arXiv Detail & Related papers (2021-06-04T17:59:57Z) - SegmentMeIfYouCan: A Benchmark for Anomaly Segmentation [111.61261419566908]
Deep neural networks (DNNs) are usually trained on a closed set of semantic classes.
They are ill-equipped to handle previously-unseen objects.
detecting and localizing such objects is crucial for safety-critical applications such as perception for automated driving.
arXiv Detail & Related papers (2021-04-30T07:58:19Z) - Meta Faster R-CNN: Towards Accurate Few-Shot Object Detection with
Attentive Feature Alignment [33.446875089255876]
Few-shot object detection (FSOD) aims to detect objects using only few examples.
We propose a meta-learning based few-shot object detection method by transferring meta-knowledge learned from data-abundant base classes to data-scarce novel classes.
arXiv Detail & Related papers (2021-04-15T19:01:27Z) - Target-Aware Object Discovery and Association for Unsupervised Video
Multi-Object Segmentation [79.6596425920849]
This paper addresses the task of unsupervised video multi-object segmentation.
We introduce a novel approach for more accurate and efficient unseen-temporal segmentation.
We evaluate the proposed approach on DAVIS$_17$ and YouTube-VIS, and the results demonstrate that it outperforms state-of-the-art methods both in segmentation accuracy and inference speed.
arXiv Detail & Related papers (2021-04-10T14:39:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.