Lifelong Ensemble Learning based on Multiple Representations for
Few-Shot Object Recognition
- URL: http://arxiv.org/abs/2205.01982v5
- Date: Tue, 9 Jan 2024 13:21:21 GMT
- Title: Lifelong Ensemble Learning based on Multiple Representations for
Few-Shot Object Recognition
- Authors: Hamidreza Kasaei, Songsong Xiong
- Abstract summary: We present a lifelong ensemble learning approach based on multiple representations to address the few-shot object recognition problem.
To facilitate lifelong learning, each approach is equipped with a memory unit for storing and retrieving object information instantly.
We have performed extensive sets of experiments to assess the performance of the proposed approach in offline, and open-ended scenarios.
- Score: 6.282068591820947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Service robots are integrating more and more into our daily lives to help us
with various tasks. In such environments, robots frequently face new objects
while working in the environment and need to learn them in an open-ended
fashion. Furthermore, such robots must be able to recognize a wide range of
object categories. In this paper, we present a lifelong ensemble learning
approach based on multiple representations to address the few-shot object
recognition problem. In particular, we form ensemble methods based on deep
representations and handcrafted 3D shape descriptors. To facilitate lifelong
learning, each approach is equipped with a memory unit for storing and
retrieving object information instantly. The proposed model is suitable for
open-ended learning scenarios where the number of 3D object categories is not
fixed and can grow over time. We have performed extensive sets of experiments
to assess the performance of the proposed approach in offline, and open-ended
scenarios. For the evaluation purpose, in addition to real object datasets, we
generate a large synthetic household objects dataset consisting of 27000 views
of 90 objects. Experimental results demonstrate the effectiveness of the
proposed method on online few-shot 3D object recognition tasks, as well as its
superior performance over the state-of-the-art open-ended learning approaches.
Furthermore, our results show that while ensemble learning is modestly
beneficial in offline settings, it is significantly beneficial in lifelong
few-shot learning situations. Additionally, we demonstrated the effectiveness
of our approach in both simulated and real-robot settings, where the robot
rapidly learned new categories from limited examples.
Related papers
- A Survey of Embodied Learning for Object-Centric Robotic Manipulation [27.569063968870868]
Embodied learning for object-centric robotic manipulation is a rapidly developing and challenging area in AI.
Unlike data-driven machine learning methods, embodied learning focuses on robot learning through physical interaction with the environment.
arXiv Detail & Related papers (2024-08-21T11:32:09Z) - Zero-Shot Object-Centric Representation Learning [72.43369950684057]
We study current object-centric methods through the lens of zero-shot generalization.
We introduce a benchmark comprising eight different synthetic and real-world datasets.
We find that training on diverse real-world images improves transferability to unseen scenarios.
arXiv Detail & Related papers (2024-08-17T10:37:07Z) - SUGAR: Pre-training 3D Visual Representations for Robotics [85.55534363501131]
We introduce a novel 3D pre-training framework for robotics named SUGAR.
SUGAR captures semantic, geometric and affordance properties of objects through 3D point clouds.
We show that SUGAR's 3D representation outperforms state-of-the-art 2D and 3D representations.
arXiv Detail & Related papers (2024-04-01T21:23:03Z) - Fine-grained 3D object recognition: an approach and experiments [0.0]
Three-dimensional (3D) object recognition technology is being used as a core technology in advanced technologies such as autonomous driving of automobiles.
There are two sets of approaches for 3D object recognition: (i) hand-crafted approaches like Global Orthographic Object Descriptor (GOOD), and (ii) deep learning-based approaches such as MobileNet and VGG.
In this paper, we first implemented an offline 3D object recognition system that takes an object view as input and generates category labels as output.
In the offline stage, instance-based learning (IBL) is used to form a new
arXiv Detail & Related papers (2023-06-28T04:48:21Z) - Visuomotor Control in Multi-Object Scenes Using Object-Aware
Representations [25.33452947179541]
We show the effectiveness of object-aware representation learning techniques for robotic tasks.
Our model learns control policies in a sample-efficient manner and outperforms state-of-the-art object techniques.
arXiv Detail & Related papers (2022-05-12T19:48:11Z) - ObjectFolder: A Dataset of Objects with Implicit Visual, Auditory, and
Tactile Representations [52.226947570070784]
We present Object, a dataset of 100 objects that addresses both challenges with two key innovations.
First, Object encodes the visual, auditory, and tactile sensory data for all objects, enabling a number of multisensory object recognition tasks.
Second, Object employs a uniform, object-centric simulations, and implicit representation for each object's visual textures, tactile readings, and tactile readings, making the dataset flexible to use and easy to share.
arXiv Detail & Related papers (2021-09-16T14:00:59Z) - RandomRooms: Unsupervised Pre-training from Synthetic Shapes and
Randomized Layouts for 3D Object Detection [138.2892824662943]
A promising solution is to make better use of the synthetic dataset, which consists of CAD object models, to boost the learning on real datasets.
Recent work on 3D pre-training exhibits failure when transfer features learned on synthetic objects to other real-world applications.
In this work, we put forward a new method called RandomRooms to accomplish this objective.
arXiv Detail & Related papers (2021-08-17T17:56:12Z) - Simultaneous Multi-View Object Recognition and Grasping in Open-Ended
Domains [0.0]
We propose a deep learning architecture with augmented memory capacities to handle open-ended object recognition and grasping simultaneously.
We demonstrate the ability of our approach to grasp never-seen-before objects and to rapidly learn new object categories using very few examples on-site in both simulation and real-world settings.
arXiv Detail & Related papers (2021-06-03T14:12:11Z) - Analysis of voxel-based 3D object detection methods efficiency for
real-time embedded systems [93.73198973454944]
Two popular voxel-based 3D object detection methods are studied in this paper.
Our experiments show that these methods mostly fail to detect distant small objects due to the sparsity of the input point clouds at large distances.
Our findings suggest that a considerable part of the computations of existing methods is focused on locations of the scene that do not contribute with successful detection.
arXiv Detail & Related papers (2021-05-21T12:40:59Z) - REGRAD: A Large-Scale Relational Grasp Dataset for Safe and
Object-Specific Robotic Grasping in Clutter [52.117388513480435]
We present a new dataset named regrad to sustain the modeling of relationships among objects and grasps.
Our dataset is collected in both forms of 2D images and 3D point clouds.
Users are free to import their own object models for the generation of as many data as they want.
arXiv Detail & Related papers (2021-04-29T05:31:21Z) - Open-Ended Fine-Grained 3D Object Categorization by Combining Shape and
Texture Features in Multiple Colorspaces [5.89118432388542]
In this work, shape information encodes the common patterns of all categories, while texture information is used to describe the appearance of each instance in detail.
The proposed network architecture out-performed the selected state-of-the-art approaches in terms of object classification accuracy and scalability.
arXiv Detail & Related papers (2020-09-19T14:06:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.