Transfer Learning for Matrix Completion
- URL: http://arxiv.org/abs/2507.02248v1
- Date: Thu, 03 Jul 2025 02:43:40 GMT
- Title: Transfer Learning for Matrix Completion
- Authors: Dali Liu, Haolei Weng,
- Abstract summary: We propose a transfer learning procedure given prior information on which source datasets are favorable.<n>With the source matrices close enough to the target matrix, out method outperforms the traditional method using the single target data.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we explore the knowledge transfer under the setting of matrix completion, which aims to enhance the estimation of a low-rank target matrix with auxiliary data available. We propose a transfer learning procedure given prior information on which source datasets are favorable. We study its convergence rates and prove its minimax optimality. Our analysis reveals that with the source matrices close enough to the target matrix, out method outperforms the traditional method using the single target data. In particular, we leverage the advanced sharp concentration inequalities introduced in \cite{brailovskaya2024universality} to eliminate a logarithmic factor in the convergence rate, which is crucial for proving the minimax optimality. When the relevance of source datasets is unknown, we develop an efficient detection procedure to identify informative sources and establish its selection consistency. Simulations and real data analysis are conducted to support the validity of our methodology.
Related papers
- Formal Bayesian Transfer Learning via the Total Risk Prior [1.8570591025615457]
We show how a particular instantiation of our prior leads to a Bayesian Lasso in a transformed coordinate system.<n>We also demonstrate that recently proposed minimax-frequentist transfer learning techniques may be viewed as an approximate Maximum a Posteriori approach to our model.
arXiv Detail & Related papers (2025-07-31T17:55:16Z) - Trans-Glasso: A Transfer Learning Approach to Precision Matrix Estimation [30.82913179485628]
We propose Trans-Glasso, a two-step transfer learning method for precision matrix estimation.
We show that Trans-Glasso achieves minimax optimality under certain conditions.
We validate Trans-Glasso in applications to gene networks across brain tissues and protein networks for various cancer subtypes.
arXiv Detail & Related papers (2024-11-23T18:30:56Z) - A Deeper Look into Second-Order Feature Aggregation for LiDAR Place Recognition [4.358456799125694]
LiDAR Place Recognition (LPR) compresses dense pointwise features into compact global descriptors.<n>First-order aggregators such as GeM and NetVLAD are widely used, but they overlook inter-feature correlations that second-order aggregation naturally captures.<n>Channel Partition-based Second-order Local Feature Aggregation (CPS) is a drop-in, partition-based second-order aggregation module that preserves all channels while producing an order-of-magnitude smaller descriptor.
arXiv Detail & Related papers (2024-09-24T09:40:22Z) - Knowledge Transfer across Multiple Principal Component Analysis Studies [8.602833477729899]
We propose a two-step transfer learning algorithm to extract useful information from multiple source principal component analysis (PCA) studies.
In the first step, we integrate the shared subspace information across multiple studies by a proposed method named as Grassmannian barycenter.
The resulting estimator for the shared subspace from the first step is further utilized to estimate the target private subspace.
arXiv Detail & Related papers (2024-03-12T09:15:12Z) - Invariance Learning in Deep Neural Networks with Differentiable Laplace
Approximations [76.82124752950148]
We develop a convenient gradient-based method for selecting the data augmentation.
We use a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective.
arXiv Detail & Related papers (2022-02-22T02:51:11Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Learning Bias-Invariant Representation by Cross-Sample Mutual
Information Minimization [77.8735802150511]
We propose a cross-sample adversarial debiasing (CSAD) method to remove the bias information misused by the target task.
The correlation measurement plays a critical role in adversarial debiasing and is conducted by a cross-sample neural mutual information estimator.
We conduct thorough experiments on publicly available datasets to validate the advantages of the proposed method over state-of-the-art approaches.
arXiv Detail & Related papers (2021-08-11T21:17:02Z) - Transfer Learning under High-dimensional Generalized Linear Models [7.675822266933702]
We study the transfer learning problem under high-dimensional generalized linear models.
We propose an oracle algorithm and derive its $ell$-estimation error bounds.
When we don't know which sources to transfer, an algorithm-free transferable source detection approach is introduced.
arXiv Detail & Related papers (2021-05-29T15:39:43Z) - Incremental Semi-Supervised Learning Through Optimal Transport [0.0]
We propose a novel approach for the transductive semi-supervised learning, using a complete bipartite edge-weighted graph.
The proposed approach uses the regularized optimal transport between empirical measures defined on labelled and unlabelled data points in order to obtain an affinity matrix from the optimal transport plan.
arXiv Detail & Related papers (2021-03-22T15:31:53Z) - Graph Embedding with Data Uncertainty [113.39838145450007]
spectral-based subspace learning is a common data preprocessing step in many machine learning pipelines.
Most subspace learning methods do not take into consideration possible measurement inaccuracies or artifacts that can lead to data with high uncertainty.
arXiv Detail & Related papers (2020-09-01T15:08:23Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.