CLEVA-Compass: A Continual Learning EValuation Assessment Compass to
Promote Research Transparency and Comparability
- URL: http://arxiv.org/abs/2110.03331v1
- Date: Thu, 7 Oct 2021 10:53:26 GMT
- Title: CLEVA-Compass: A Continual Learning EValuation Assessment Compass to
Promote Research Transparency and Comparability
- Authors: Martin Mundt, Steven Lang, Quentin Delfosse, Kristian Kersting
- Abstract summary: We argue that the goal of a precise formulation of desiderata is an ill-posed one, as diverse applications may always warrant distinct scenarios.
In addition to promoting compact specification in the spirit of recent replication trends, the CLEVA- compass provides an intuitive chart to understand the priorities of individual systems.
- Score: 15.342039156426843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: What is the state of the art in continual machine learning? Although a
natural question for predominant static benchmarks, the notion to train systems
in a lifelong manner entails a plethora of additional challenges with respect
to set-up and evaluation. The latter have recently sparked a growing amount of
critiques on prominent algorithm-centric perspectives and evaluation protocols
being too narrow, resulting in several attempts at constructing guidelines in
favor of specific desiderata or arguing against the validity of prevalent
assumptions. In this work, we depart from this mindset and argue that the goal
of a precise formulation of desiderata is an ill-posed one, as diverse
applications may always warrant distinct scenarios. Instead, we introduce the
Continual Learning EValuation Assessment Compass, CLEVA-Compass for short. The
compass provides the visual means to both identify how approaches are
practically reported and how works can simultaneously be contextualized in the
broader literature landscape. In addition to promoting compact specification in
the spirit of recent replication trends, the CLEVA-Compass thus provides an
intuitive chart to understand the priorities of individual systems, where they
resemble each other, and what elements are missing towards a fair comparison.
Related papers
- Towards Distribution-Agnostic Generalized Category Discovery [51.52673017664908]
Data imbalance and open-ended distribution are intrinsic characteristics of the real visual world.
We propose a Self-Balanced Co-Advice contrastive framework (BaCon)
BaCon consists of a contrastive-learning branch and a pseudo-labeling branch, working collaboratively to provide interactive supervision to resolve the DA-GCD task.
arXiv Detail & Related papers (2023-10-02T17:39:58Z) - Large Class Separation is not what you need for Relational
Reasoning-based OOD Detection [12.578844450586]
Out-Of-Distribution (OOD) detection methods provide a solution by identifying semantic novelty.
Most of these methods leverage a learning stage on the known data, which means training (or fine-tuning) a model to capture the concept of normality.
A viable alternative is that of evaluating similarities in the embedding space produced by large pre-trained models without any further learning effort.
arXiv Detail & Related papers (2023-07-12T14:10:15Z) - Fairness meets Cross-Domain Learning: a new perspective on Models and
Metrics [80.07271410743806]
We study the relationship between cross-domain learning (CD) and model fairness.
We introduce a benchmark on face and medical images spanning several demographic groups as well as classification and localization tasks.
Our study covers 14 CD approaches alongside three state-of-the-art fairness algorithms and shows how the former can outperform the latter.
arXiv Detail & Related papers (2023-03-25T09:34:05Z) - A Low Rank Promoting Prior for Unsupervised Contrastive Learning [108.91406719395417]
We construct a novel probabilistic graphical model that effectively incorporates the low rank promoting prior into the framework of contrastive learning.
Our hypothesis explicitly requires that all the samples belonging to the same instance class lie on the same subspace with small dimension.
Empirical evidences show that the proposed algorithm clearly surpasses the state-of-the-art approaches on multiple benchmarks.
arXiv Detail & Related papers (2021-08-05T15:58:25Z) - Self-Point-Flow: Self-Supervised Scene Flow Estimation from Point Clouds
with Optimal Transport and Random Walk [59.87525177207915]
We develop a self-supervised method to establish correspondences between two point clouds to approximate scene flow.
Our method achieves state-of-the-art performance among self-supervised learning methods.
arXiv Detail & Related papers (2021-05-18T03:12:42Z) - Learning with Instance Bundles for Reading Comprehension [61.823444215188296]
We introduce new supervision techniques that compare question-answer scores across multiple related instances.
Specifically, we normalize these scores across various neighborhoods of closely contrasting questions and/or answers.
We empirically demonstrate the effectiveness of training with instance bundles on two datasets.
arXiv Detail & Related papers (2021-04-18T06:17:54Z) - A Survey on Contrastive Self-supervised Learning [0.0]
Self-supervised learning has gained popularity because of its ability to avoid the cost of annotating large-scale datasets.
Contrastive learning has recently become a dominant component in self-supervised learning methods for computer vision, natural language processing (NLP), and other domains.
This paper provides an extensive review of self-supervised methods that follow the contrastive approach.
arXiv Detail & Related papers (2020-10-31T21:05:04Z) - Robust Learning Through Cross-Task Consistency [92.42534246652062]
We propose a broadly applicable and fully computational method for augmenting learning with Cross-Task Consistency.
We observe that learning with cross-task consistency leads to more accurate predictions and better generalization to out-of-distribution inputs.
arXiv Detail & Related papers (2020-06-07T09:24:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.