ORBIT: A Real-World Few-Shot Dataset for Teachable Object Recognition
- URL: http://arxiv.org/abs/2104.03841v2
- Date: Fri, 9 Apr 2021 16:56:43 GMT
- Title: ORBIT: A Real-World Few-Shot Dataset for Teachable Object Recognition
- Authors: Daniela Massiceti, Luisa Zintgraf, John Bronskill, Lida Theodorou,
Matthew Tobias Harris, Edward Cutrell, Cecily Morrison, Katja Hofmann, Simone
Stumpf
- Abstract summary: We present the ORBIT dataset and benchmark, grounded in a real-world application of teachable object recognizers for people who are blind/low vision.
The dataset contains 3,822 videos of 486 objects recorded by people who are blind/low-vision on their mobile phones.
The benchmark reflects a realistic, highly challenging recognition problem, providing a rich playground to drive research in robustness to few-shot, high-variation conditions.
- Score: 21.594641488685376
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Object recognition has made great advances in the last decade, but
predominately still relies on many high-quality training examples per object
category. In contrast, learning new objects from only a few examples could
enable many impactful applications from robotics to user personalization. Most
few-shot learning research, however, has been driven by benchmark datasets that
lack the high variation that these applications will face when deployed in the
real-world. To close this gap, we present the ORBIT dataset and benchmark,
grounded in a real-world application of teachable object recognizers for people
who are blind/low vision. The dataset contains 3,822 videos of 486 objects
recorded by people who are blind/low-vision on their mobile phones, and the
benchmark reflects a realistic, highly challenging recognition problem,
providing a rich playground to drive research in robustness to few-shot,
high-variation conditions. We set the first state-of-the-art on the benchmark
and show that there is massive scope for further innovation, holding the
potential to impact a broad range of real-world vision applications including
tools for the blind/low-vision community. The dataset is available at
https://bit.ly/2OyElCj and the code to run the benchmark at
https://bit.ly/39YgiUW.
Related papers
- Plain-Det: A Plain Multi-Dataset Object Detector [22.848784430833835]
Plain-Det offers flexibility to accommodate new datasets, in performance across diverse datasets, and training efficiency.
We conduct extensive experiments on 13 downstream datasets and Plain-Det demonstrates strong generalization capability.
arXiv Detail & Related papers (2024-07-14T05:18:06Z) - Deep Learning-Based Object Pose Estimation: A Comprehensive Survey [73.74933379151419]
We discuss the recent advances in deep learning-based object pose estimation.
Our survey also covers multiple input data modalities, degrees-of-freedom of output poses, object properties, and downstream tasks.
arXiv Detail & Related papers (2024-05-13T14:44:22Z) - Open World Object Detection in the Era of Foundation Models [53.683963161370585]
We introduce a new benchmark that includes five real-world application-driven datasets.
We introduce a novel method, Foundation Object detection Model for the Open world, or FOMO, which identifies unknown objects based on their shared attributes with the base known objects.
arXiv Detail & Related papers (2023-12-10T03:56:06Z) - The ObjectFolder Benchmark: Multisensory Learning with Neural and Real
Objects [51.22194706674366]
We introduce the Object Benchmark, a benchmark suite of 10 tasks for multisensory object-centric learning.
We also introduce the Object Real dataset, including the multisensory measurements for 100 real-world household objects.
arXiv Detail & Related papers (2023-06-01T17:51:22Z) - OVTrack: Open-Vocabulary Multiple Object Tracking [64.73379741435255]
OVTrack is an open-vocabulary tracker capable of tracking arbitrary object classes.
It sets a new state-of-the-art on the large-scale, large-vocabulary TAO benchmark.
arXiv Detail & Related papers (2023-04-17T16:20:05Z) - Lifelong Ensemble Learning based on Multiple Representations for
Few-Shot Object Recognition [6.282068591820947]
We present a lifelong ensemble learning approach based on multiple representations to address the few-shot object recognition problem.
To facilitate lifelong learning, each approach is equipped with a memory unit for storing and retrieving object information instantly.
We have performed extensive sets of experiments to assess the performance of the proposed approach in offline, and open-ended scenarios.
arXiv Detail & Related papers (2022-05-04T10:29:10Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - ObjectFolder: A Dataset of Objects with Implicit Visual, Auditory, and
Tactile Representations [52.226947570070784]
We present Object, a dataset of 100 objects that addresses both challenges with two key innovations.
First, Object encodes the visual, auditory, and tactile sensory data for all objects, enabling a number of multisensory object recognition tasks.
Second, Object employs a uniform, object-centric simulations, and implicit representation for each object's visual textures, tactile readings, and tactile readings, making the dataset flexible to use and easy to share.
arXiv Detail & Related papers (2021-09-16T14:00:59Z) - F-SIOL-310: A Robotic Dataset and Benchmark for Few-Shot Incremental
Object Learning [9.89901717499058]
We present F-SIOL-310 (Few-Shot Incremental Object Learning) for testing few-shot incremental object learning capability for robotic vision.
We also provide benchmarks and evaluations of 8 incremental learning algorithms on F-SIOL-310 for future comparisons.
arXiv Detail & Related papers (2021-03-23T00:25:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.