Multi-Dictionary Tensor Decomposition
- URL: http://arxiv.org/abs/2309.09717v1
- Date: Mon, 18 Sep 2023 12:31:56 GMT
- Title: Multi-Dictionary Tensor Decomposition
- Authors: Maxwell McNeil and Petko Bogdanov
- Abstract summary: We propose a framework for Multi-Dictionary Decomposition (MDTD)
We derive a general optimization algorithm for MDTD that handles both complete input and input with missing values.
It can impute missing values in billion-entry tensors more accurately and scalably than state-of-the-art competitors.
- Score: 5.733331864416094
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Tensor decomposition methods are popular tools for analysis of multi-way
datasets from social media, healthcare, spatio-temporal domains, and others.
Widely adopted models such as Tucker and canonical polyadic decomposition (CPD)
follow a data-driven philosophy: they decompose a tensor into factors that
approximate the observed data well. In some cases side information is available
about the tensor modes. For example, in a temporal user-item purchases tensor a
user influence graph, an item similarity graph, and knowledge about seasonality
or trends in the temporal mode may be available. Such side information may
enable more succinct and interpretable tensor decomposition models and improved
quality in downstream tasks.
We propose a framework for Multi-Dictionary Tensor Decomposition (MDTD) which
takes advantage of prior structural information about tensor modes in the form
of coding dictionaries to obtain sparsely encoded tensor factors. We derive a
general optimization algorithm for MDTD that handles both complete input and
input with missing values. Our framework handles large sparse tensors typical
to many real-world application domains. We demonstrate MDTD's utility via
experiments with both synthetic and real-world datasets. It learns more concise
models than dictionary-free counterparts and improves (i) reconstruction
quality ($60\%$ fewer non-zero coefficients coupled with smaller error); (ii)
missing values imputation quality (two-fold MSE reduction with up to orders of
magnitude time savings) and (iii) the estimation of the tensor rank. MDTD's
quality improvements do not come with a running time premium: it can decompose
$19GB$ datasets in less than a minute. It can also impute missing values in
sparse billion-entry tensors more accurately and scalably than state-of-the-art
competitors.
Related papers
- Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - FRAPPE: $\underline{\text{F}}$ast $\underline{\text{Ra}}$nk $\underline{\text{App}}$roximation with $\underline{\text{E}}$xplainable Features for Tensors [5.39764619690516]
FRAPPE is the first method to estimate the canonical rank of a tensor without having to compute the CPD.
It is over 24 times faster than the best-performing baseline and exhibits a 10% improvement in MAPE on a synthetic dataset.
arXiv Detail & Related papers (2022-06-19T03:19:59Z) - Truncated tensor Schatten p-norm based approach for spatiotemporal
traffic data imputation with complicated missing patterns [77.34726150561087]
We introduce four complicated missing patterns, including missing and three fiber-like missing cases according to the mode-drivenn fibers.
Despite nonity of the objective function in our model, we derive the optimal solutions by integrating alternating data-mputation method of multipliers.
arXiv Detail & Related papers (2022-05-19T08:37:56Z) - Augmented Tensor Decomposition with Stochastic Optimization [46.16865811396394]
Real-world tensor data are usually high-ordered and have large dimensions with millions or billions of entries.
It is expensive to decompose the whole tensor with traditional algorithms.
This paper proposes augmented tensor decomposition, which effectively incorporates data augmentations to boost downstream classification.
arXiv Detail & Related papers (2021-06-15T06:29:05Z) - MTC: Multiresolution Tensor Completion from Partial and Coarse
Observations [49.931849672492305]
Existing completion formulation mostly relies on partial observations from a single tensor.
We propose an efficient Multi-resolution Completion model (MTC) to solve the problem.
arXiv Detail & Related papers (2021-06-14T02:20:03Z) - Cherry-Picking Gradients: Learning Low-Rank Embeddings of Visual Data
via Differentiable Cross-Approximation [53.95297550117153]
We propose an end-to-end trainable framework that processes large-scale visual data tensors by looking emphat a fraction of their entries only.
The proposed approach is particularly useful for large-scale multidimensional grid data, and for tasks that require context over a large receptive field.
arXiv Detail & Related papers (2021-05-29T08:39:57Z) - Multi-version Tensor Completion for Time-delayed Spatio-temporal Data [50.762087239885936]
Real-world-temporal data is often incomplete or inaccurate due to various data loading delays.
We propose a low-rank tensor model to predict the updates over time.
We obtain up to 27.2% lower root mean-squared-error compared to the best baseline method.
arXiv Detail & Related papers (2021-05-11T19:55:56Z) - Low-Rank and Sparse Enhanced Tucker Decomposition for Tensor Completion [3.498620439731324]
We introduce a unified low-rank and sparse enhanced Tucker decomposition model for tensor completion.
Our model possesses a sparse regularization term to promote a sparse core tensor, which is beneficial for tensor data compression.
It is remarkable that our model is able to deal with different types of real-world data sets, since it exploits the potential periodicity and inherent correlation properties appeared in tensors.
arXiv Detail & Related papers (2020-10-01T12:45:39Z) - Sparse and Low-Rank High-Order Tensor Regression via Parallel Proximal
Method [6.381138694845438]
We propose the Sparse and Low-rank Regression model for large-scale data with high-order structures.
Our model enforces sparsity and low-rankness of the tensor coefficient.
Our model's predictions exhibit meaningful interpretations on the video dataset.
arXiv Detail & Related papers (2019-11-29T06:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.