PCA-based Multi Task Learning: a Random Matrix Approach
- URL: http://arxiv.org/abs/2111.00924v1
- Date: Mon, 1 Nov 2021 13:13:38 GMT
- Title: PCA-based Multi Task Learning: a Random Matrix Approach
- Authors: Malik Tiomoko, Romain Couillet and Fr\'ed\'eric Pascal
- Abstract summary: The article proposes and theoretically analyses a emphcomputationally efficient multi-task learning (MTL) extension of popular principal component analysis (PCA)-based supervised learning schemes citebarshan2011supervised,bair2006prediction.
- Score: 40.49988553835459
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The article proposes and theoretically analyses a \emph{computationally
efficient} multi-task learning (MTL) extension of popular principal component
analysis (PCA)-based supervised learning schemes
\cite{barshan2011supervised,bair2006prediction}. The analysis reveals that (i)
by default learning may dramatically fail by suffering from \emph{negative
transfer}, but that (ii) simple counter-measures on data labels avert negative
transfer and necessarily result in improved performances.
Supporting experiments on synthetic and real data benchmarks show that the
proposed method achieves comparable performance with state-of-the-art MTL
methods but at a \emph{significantly reduced computational cost}.
Related papers
- Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Rethinking Collaborative Metric Learning: Toward an Efficient
Alternative without Negative Sampling [156.7248383178991]
Collaborative Metric Learning (CML) paradigm has aroused wide interest in the area of recommendation systems (RS)
We find that negative sampling would lead to a biased estimation of the generalization error.
Motivated by this, we propose an efficient alternative without negative sampling for CML named textitSampling-Free Collaborative Metric Learning (SFCML)
arXiv Detail & Related papers (2022-06-23T08:50:22Z) - DEALIO: Data-Efficient Adversarial Learning for Imitation from
Observation [57.358212277226315]
In imitation learning from observation IfO, a learning agent seeks to imitate a demonstrating agent using only observations of the demonstrated behavior without access to the control signals generated by the demonstrator.
Recent methods based on adversarial imitation learning have led to state-of-the-art performance on IfO problems, but they typically suffer from high sample complexity due to a reliance on data-inefficient, model-free reinforcement learning algorithms.
This issue makes them impractical to deploy in real-world settings, where gathering samples can incur high costs in terms of time, energy, and risk.
We propose a more data-efficient IfO algorithm
arXiv Detail & Related papers (2021-03-31T23:46:32Z) - Robust Principal Component Analysis: A Median of Means Approach [17.446104539598895]
Principal Component Analysis is a tool for data visualization, denoising, and dimensionality reduction.
Recent supervised learning methods have shown great success in dealing with outlying observations.
This paper proposes a PCA procedure based on the MoM principle.
arXiv Detail & Related papers (2021-02-05T19:59:05Z) - Robust and flexible learning of a high-dimensional classification rule
using auxiliary outcomes [2.92281985958308]
We develop a transfer learning approach to estimating a high-dimensional linear decision rule with the presence of auxiliary outcomes.
We show that the final estimator can achieve a lower estimation error than the one using only the single outcome of interest.
arXiv Detail & Related papers (2020-11-11T01:14:33Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z) - Large Dimensional Analysis and Improvement of Multi Task Learning [38.86699890656948]
Multi Task Learning (MTL) efficiently leverages useful information contained in multiple related tasks to help improve the generalization performance of all tasks.
This article conducts a large dimensional analysis of a simple but, as we shall see, extremely powerful when carefully tuned, Least Square Support Vector Machine (LSSVM) version of MTL.
arXiv Detail & Related papers (2020-09-03T11:40:14Z) - Modal Principal Component Analysis [3.050919759387985]
It has been shown that the robustness of many statistical methods can be improved using mode estimation instead of mean estimation.
This study proposes a modal principal component analysis (MPCA) which is a robust PCA method based on mode estimation.
arXiv Detail & Related papers (2020-08-07T23:59:05Z) - Prototypical Contrastive Learning of Unsupervised Representations [171.3046900127166]
Prototypical Contrastive Learning (PCL) is an unsupervised representation learning method.
PCL implicitly encodes semantic structures of the data into the learned embedding space.
PCL outperforms state-of-the-art instance-wise contrastive learning methods on multiple benchmarks.
arXiv Detail & Related papers (2020-05-11T09:53:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.