Contrastive Learning for Cross-Domain Open World Recognition
- URL: http://arxiv.org/abs/2203.09257v1
- Date: Thu, 17 Mar 2022 11:23:53 GMT
- Title: Contrastive Learning for Cross-Domain Open World Recognition
- Authors: Francesco Cappio Borlino, Silvia Bucci, Tatiana Tommasi
- Abstract summary: The ability to evolve is fundamental for any valuable autonomous agent whose knowledge cannot remain limited to that injected by the manufacturer.
We show how it learns a feature space perfectly suitable to incrementally include new classes and is able to capture knowledge which generalizes across a variety of visual domains.
Our method is endowed with a tailored effective stopping criterion for each learning episode and exploits a novel self-paced thresholding strategy.
- Score: 17.660958043781154
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ability to evolve is fundamental for any valuable autonomous agent whose
knowledge cannot remain limited to that injected by the manufacturer. Consider
for example a home assistant robot: it should be able to incrementally learn
new object categories when requested, but also to recognize the same objects in
different environments (rooms) and poses (hand-held/on the floor/above
furniture), while rejecting unknown ones. Despite its importance, this scenario
has started to raise interest in the robotic community only recently and the
related research is still in its infancy, with existing experimental testbeds
but no tailored methods. With this work, we propose the first learning approach
that deals with all the previously mentioned challenges at once by exploiting a
single contrastive objective. We show how it learns a feature space perfectly
suitable to incrementally include new classes and is able to capture knowledge
which generalizes across a variety of visual domains. Our method is endowed
with a tailored effective stopping criterion for each learning episode and
exploits a novel self-paced thresholding strategy that provides the classifier
with a reliable rejection option. Both these contributions are based on the
observation of the data statistics and do not need manual tuning. An extensive
experimental analysis confirms the effectiveness of the proposed approach
establishing the new state-of-the-art. The code is available at
https://github.com/FrancescoCappio/Contrastive_Open_World.
Related papers
- ConceptDrift: Uncovering Biases through the Lens of Foundation Models [5.025665239455297]
We show that ConceptDrift can be scaled to automatically identify biases in datasets without human prior knowledge.
We propose two bias identification evaluation protocols to fill the gap in the previous work and show that our method significantly improves over SoTA methods.
Our method is not bounded to a single modality, and we empirically validate it both on image (Waterbirds, CelebA, ImageNet, and text datasets (CivilComments))
arXiv Detail & Related papers (2024-10-24T17:59:16Z) - Class incremental learning with probability dampening and cascaded gated classifier [4.285597067389559]
We propose a novel incremental regularisation approach called Margin Dampening and Cascaded Scaling.
The first combines a soft constraint and a knowledge distillation approach to preserve past knowledge while allowing forgetting new patterns.
We empirically show that our approach performs well on multiple benchmarks well-established baselines.
arXiv Detail & Related papers (2024-02-02T09:33:07Z) - CDFSL-V: Cross-Domain Few-Shot Learning for Videos [58.37446811360741]
Few-shot video action recognition is an effective approach to recognizing new categories with only a few labeled examples.
Existing methods in video action recognition rely on large labeled datasets from the same domain.
We propose a novel cross-domain few-shot video action recognition method that leverages self-supervised learning and curriculum learning.
arXiv Detail & Related papers (2023-09-07T19:44:27Z) - Learning Classifiers of Prototypes and Reciprocal Points for Universal
Domain Adaptation [79.62038105814658]
Universal Domain aims to transfer the knowledge between datasets by handling two shifts: domain-shift and categoryshift.
Main challenge is correctly distinguishing the unknown target samples while adapting the distribution of known class knowledge from source to target.
Most existing methods approach this problem by first training the target adapted known and then relying on the single threshold to distinguish unknown target samples.
arXiv Detail & Related papers (2022-12-16T09:01:57Z) - Adversarial Auto-Augment with Label Preservation: A Representation
Learning Principle Guided Approach [95.74102207187545]
We show that a prior-free autonomous data augmentation's objective can be derived from a representation learning principle.
We then propose a practical surrogate to the objective that can be efficiently optimized and integrated seamlessly into existing methods.
arXiv Detail & Related papers (2022-11-02T02:02:51Z) - Novel Class Discovery without Forgetting [72.52222295216062]
We identify and formulate a new, pragmatic problem setting of NCDwF: Novel Class Discovery without Forgetting.
We propose a machine learning model to incrementally discover novel categories of instances from unlabeled data.
We introduce experimental protocols based on CIFAR-10, CIFAR-100 and ImageNet-1000 to measure the trade-off between knowledge retention and novel class discovery.
arXiv Detail & Related papers (2022-07-21T17:54:36Z) - Semantic Novelty Detection via Relational Reasoning [17.660958043781154]
We propose a novel representation learning paradigm based on relational reasoning.
Our experiments show that this knowledge is directly transferable to a wide range of scenarios.
It can be exploited as a plug-and-play module to convert closed-set recognition models into reliable open-set ones.
arXiv Detail & Related papers (2022-07-18T15:49:27Z) - On Generalizing Beyond Domains in Cross-Domain Continual Learning [91.56748415975683]
Deep neural networks often suffer from catastrophic forgetting of previously learned knowledge after learning a new task.
Our proposed approach learns new tasks under domain shift with accuracy boosts up to 10% on challenging datasets such as DomainNet and OfficeHome.
arXiv Detail & Related papers (2022-03-08T09:57:48Z) - A Survey on Contrastive Self-supervised Learning [0.0]
Self-supervised learning has gained popularity because of its ability to avoid the cost of annotating large-scale datasets.
Contrastive learning has recently become a dominant component in self-supervised learning methods for computer vision, natural language processing (NLP), and other domains.
This paper provides an extensive review of self-supervised methods that follow the contrastive approach.
arXiv Detail & Related papers (2020-10-31T21:05:04Z) - COG: Connecting New Skills to Past Experience with Offline Reinforcement
Learning [78.13740204156858]
We show that we can reuse prior data to extend new skills simply through dynamic programming.
We demonstrate the effectiveness of our approach by chaining together several behaviors seen in prior datasets for solving a new task.
We train our policies in an end-to-end fashion, mapping high-dimensional image observations to low-level robot control commands.
arXiv Detail & Related papers (2020-10-27T17:57:29Z) - Self-Supervised Learning Aided Class-Incremental Lifelong Learning [17.151579393716958]
We study the issue of catastrophic forgetting in class-incremental learning (Class-IL)
In training procedure of Class-IL, as the model has no knowledge about following tasks, it would only extract features necessary for tasks learned so far, whose information is insufficient for joint classification.
We propose to combine self-supervised learning, which can provide effective representations without requiring labels, with Class-IL to partly get around this problem.
arXiv Detail & Related papers (2020-06-10T15:15:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.