Analysis of Self-Supervised Learning and Dimensionality Reduction
Methods in Clustering-Based Active Learning for Speech Emotion Recognition
- URL: http://arxiv.org/abs/2206.10188v1
- Date: Tue, 21 Jun 2022 08:44:55 GMT
- Title: Analysis of Self-Supervised Learning and Dimensionality Reduction
Methods in Clustering-Based Active Learning for Speech Emotion Recognition
- Authors: Einari Vaaras, Manu Airaksinen, Okko R\"as\"anen
- Abstract summary: We show how to use the structure of the feature space for clustering-based active learning (AL) methods.
In this paper, we combine CPC and multiple dimensionality reduction methods in search of functioning practices for clustering-based AL.
Our experiments for simulating speech emotion recognition system deployment show that both the local and global topology of the feature space can be successfully used for AL.
- Score: 3.3670613441132984
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When domain experts are needed to perform data annotation for complex
machine-learning tasks, reducing annotation effort is crucial in order to cut
down time and expenses. For cases when there are no annotations available, one
approach is to utilize the structure of the feature space for clustering-based
active learning (AL) methods. However, these methods are heavily dependent on
how the samples are organized in the feature space and what distance metric is
used. Unsupervised methods such as contrastive predictive coding (CPC) can
potentially be used to learn organized feature spaces, but these methods
typically create high-dimensional features which might be challenging for
estimating data density. In this paper, we combine CPC and multiple
dimensionality reduction methods in search of functioning practices for
clustering-based AL. Our experiments for simulating speech emotion recognition
system deployment show that both the local and global topology of the feature
space can be successfully used for AL, and that CPC can be used to improve
clustering-based AL performance over traditional signal features. Additionally,
we observe that compressing data dimensionality does not harm AL performance
substantially, and that 2-D feature representations achieved similar AL
performance as higher-dimensional representations when the number of
annotations is not very low.
Related papers
- Beyond Prototypes: Semantic Anchor Regularization for Better
Representation Learning [82.29761875805369]
One of the ultimate goals of representation learning is to achieve compactness within a class and well-separability between classes.
We propose a novel perspective to use pre-defined class anchors serving as feature centroid to unidirectionally guide feature learning.
The proposed Semantic Anchor Regularization (SAR) can be used in a plug-and-play manner in the existing models.
arXiv Detail & Related papers (2023-12-19T05:52:38Z) - Learning Common Rationale to Improve Self-Supervised Representation for
Fine-Grained Visual Recognition Problems [61.11799513362704]
We propose learning an additional screening mechanism to identify discriminative clues commonly seen across instances and classes.
We show that a common rationale detector can be learned by simply exploiting the GradCAM induced from the SSL objective.
arXiv Detail & Related papers (2023-03-03T02:07:40Z) - Hyperspherical Consistency Regularization [45.00073340936437]
We explore the relationship between self-supervised learning and supervised learning, and study how self-supervised learning helps robust data-efficient deep learning.
We propose hyperspherical consistency regularization (HCR), a simple yet effective plug-and-play method, to regularize the classifier using feature-dependent information and thus avoid bias from labels.
arXiv Detail & Related papers (2022-06-02T02:41:13Z) - Spacing Loss for Discovering Novel Categories [72.52222295216062]
Novel Class Discovery (NCD) is a learning paradigm, where a machine learning model is tasked to semantically group instances from unlabeled data.
We first characterize existing NCD approaches into single-stage and two-stage methods based on whether they require access to labeled and unlabeled data together.
We devise a simple yet powerful loss function that enforces separability in the latent space using cues from multi-dimensional scaling.
arXiv Detail & Related papers (2022-04-22T09:37:11Z) - Clustering augmented Self-Supervised Learning: Anapplication to Land
Cover Mapping [10.720852987343896]
We introduce a new method for land cover mapping by using a clustering based pretext task for self-supervised learning.
We demonstrate the effectiveness of the method on two societally relevant applications.
arXiv Detail & Related papers (2021-08-16T19:35:43Z) - Combining Feature and Instance Attribution to Detect Artifacts [62.63504976810927]
We propose methods to facilitate identification of training data artifacts.
We show that this proposed training-feature attribution approach can be used to uncover artifacts in training data.
We execute a small user study to evaluate whether these methods are useful to NLP researchers in practice.
arXiv Detail & Related papers (2021-07-01T09:26:13Z) - A Framework to Enhance Generalization of Deep Metric Learning methods
using General Discriminative Feature Learning and Class Adversarial Neural
Networks [1.5469452301122175]
Metric learning algorithms aim to learn a distance function that brings semantically similar data items together and keeps dissimilar ones at a distance.
Deep Metric Learning (DML) methods are proposed that automatically extract features from data and learn a non-linear transformation from input space to a semantically embedding space.
We propose a framework to enhance the generalization power of existing DML methods in a Zero-Shot Learning (ZSL) setting.
arXiv Detail & Related papers (2021-06-11T14:24:40Z) - ReliefE: Feature Ranking in High-dimensional Spaces via Manifold
Embeddings [0.0]
Relief family of algorithms assign importances to features by iteratively accounting for nearest relevant and irrelevant instances.
Recent embedding-based methods learn compact, low-dimensional representations.
ReliefE algorithm is faster and can result in better feature rankings.
arXiv Detail & Related papers (2021-01-23T20:23:31Z) - Fantastic Features and Where to Find Them: Detecting Cognitive
Impairment with a Subsequence Classification Guided Approach [6.063165888023164]
We describe a new approach to feature engineering that leverages sequential machine learning models and domain knowledge to predict which features help enhance performance.
We demonstrate that CI classification accuracy improves by 2.3% over a strong baseline when using features produced by this method.
arXiv Detail & Related papers (2020-10-13T17:57:18Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z) - Pairwise Similarity Knowledge Transfer for Weakly Supervised Object
Localization [53.99850033746663]
We study the problem of learning localization model on target classes with weakly supervised image labels.
In this work, we argue that learning only an objectness function is a weak form of knowledge transfer.
Experiments on the COCO and ILSVRC 2013 detection datasets show that the performance of the localization model improves significantly with the inclusion of pairwise similarity function.
arXiv Detail & Related papers (2020-03-18T17:53:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.