Orthogonal Inductive Matrix Completion
- URL: http://arxiv.org/abs/2004.01653v6
- Date: Wed, 25 Aug 2021 13:31:52 GMT
- Title: Orthogonal Inductive Matrix Completion
- Authors: Antoine Ledent, Rodrigo Alves, and Marius Kloft
- Abstract summary: We propose an interpretable approach to matrix completion based on a sum of orthonormal side information terms.
We optimize the approach by a provably converging algorithm.
We analyse the performance of OMIC on several synthetic and real datasets.
- Score: 25.03115399173275
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose orthogonal inductive matrix completion (OMIC), an interpretable
approach to matrix completion based on a sum of multiple orthonormal side
information terms, together with nuclear-norm regularization.
The approach allows us to inject prior knowledge about the singular vectors
of the ground truth matrix.
We optimize the approach by a provably converging algorithm, which optimizes
all components of the model simultaneously. We study the generalization
capabilities of our method in both the distribution-free setting and in the
case where the sampling distribution admits uniform marginals, yielding
learning guarantees that improve with the quality of the injected knowledge in
both cases. As particular cases of our framework, we present models which can
incorporate user and item biases or community information in a joint and
additive fashion.
We analyse the performance of OMIC on several synthetic and real datasets.
On synthetic datasets with a sliding scale of user bias relevance, we show
that OMIC better adapts to different regimes than other methods. On real-life
datasets containing user/items recommendations and relevant side information,
we find that OMIC surpasses the state-of-the-art, with the added benefit of
greater interpretability.
Related papers
- Nonparametric Linear Discriminant Analysis for High Dimensional Matrix-Valued Data [0.0]
We propose a novel extension of Fisher's Linear Discriminant Analysis (LDA) tailored for matrix-valued observations.<n>We adopt a nonparametric empirical Bayes framework based on Non Maximum Likelihood Estimation (NPMLE)<n>Our method is effectively generalized to the matrix setting, thereby improving classification performance.
arXiv Detail & Related papers (2025-07-25T07:30:24Z) - Combining Entropy and Matrix Nuclear Norm for Enhanced Evaluation of Language Models [0.0]
Large language models (LLMs) continue to advance, the need for precise and efficient evaluation metrics becomes more pressing.
Traditional approaches, while informative, often face limitations in computational demands and interpretability.
In this paper, we introduce a novel hybrid evaluation method that integrates two established techniques.
arXiv Detail & Related papers (2024-10-18T14:03:52Z) - Synergistic eigenanalysis of covariance and Hessian matrices for enhanced binary classification [72.77513633290056]
We present a novel approach that combines the eigenanalysis of a covariance matrix evaluated on a training set with a Hessian matrix evaluated on a deep learning model.
Our method captures intricate patterns and relationships, enhancing classification performance.
arXiv Detail & Related papers (2024-02-14T16:10:42Z) - Self-Supervised Dataset Distillation for Transfer Learning [77.4714995131992]
We propose a novel problem of distilling an unlabeled dataset into a set of small synthetic samples for efficient self-supervised learning (SSL)
We first prove that a gradient of synthetic samples with respect to a SSL objective in naive bilevel optimization is textitbiased due to randomness originating from data augmentations or masking.
We empirically validate the effectiveness of our method on various applications involving transfer learning.
arXiv Detail & Related papers (2023-10-10T10:48:52Z) - Estimate-Then-Optimize versus Integrated-Estimation-Optimization versus
Sample Average Approximation: A Stochastic Dominance Perspective [15.832111591654293]
We show that a reverse behavior appears when the model class is well-specified and there is sufficient data.
We also demonstrate how standard sample average approximation (SAA) performs the worst when the model class is well-specified in terms of regret.
arXiv Detail & Related papers (2023-04-13T21:54:53Z) - Greedy Modality Selection via Approximate Submodular Maximization [19.22947539760366]
Multimodal learning considers learning from multi-modality data, aiming to fuse heterogeneous sources of information.
It is not always feasible to leverage all available modalities due to memory constraints.
We study modality selection, intending to efficiently select the most informative and complementary modalities under certain computational constraints.
arXiv Detail & Related papers (2022-10-22T22:07:27Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Learning Multimodal VAEs through Mutual Supervision [72.77685889312889]
MEME combines information between modalities implicitly through mutual supervision.
We demonstrate that MEME outperforms baselines on standard metrics across both partial and complete observation schemes.
arXiv Detail & Related papers (2021-06-23T17:54:35Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z) - Asymptotic Analysis of an Ensemble of Randomly Projected Linear
Discriminants [94.46276668068327]
In [1], an ensemble of randomly projected linear discriminants is used to classify datasets.
We develop a consistent estimator of the misclassification probability as an alternative to the computationally-costly cross-validation estimator.
We also demonstrate the use of our estimator for tuning the projection dimension on both real and synthetic data.
arXiv Detail & Related papers (2020-04-17T12:47:04Z) - Discrete-Valued Latent Preference Matrix Estimation with Graph Side
Information [12.836994708337144]
We develop an algorithm that matches the optimal sample complexity.
Our algorithm is robust to model errors and outperforms the existing algorithms in terms of prediction performance.
arXiv Detail & Related papers (2020-03-16T06:29:24Z) - Expected Information Maximization: Using the I-Projection for Mixture
Density Estimation [22.096148237257644]
Modelling highly multi-modal data is a challenging problem in machine learning.
We present a new algorithm called Expected Information Maximization (EIM) for computing the I-projection.
We show that our algorithm is much more effective in computing the I-projection than recent GAN approaches.
arXiv Detail & Related papers (2020-01-23T17:24:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.