Streaming Probabilistic Deep Tensor Factorization
- URL: http://arxiv.org/abs/2007.07367v1
- Date: Tue, 14 Jul 2020 21:25:39 GMT
- Title: Streaming Probabilistic Deep Tensor Factorization
- Authors: Shikai Fang, Zheng Wang, Zhimeng Pan, Ji Liu, Shandian Zhe
- Abstract summary: We propose SPIDER, a Streaming ProbabilistIc Deep tEnsoR factorization method.
We develop an efficient streaming posterior inference algorithm in the assumed-density-filtering and expectation propagation framework.
We show the advantages of our approach in four real-world applications.
- Score: 27.58928876734886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the success of existing tensor factorization methods, most of them
conduct a multilinear decomposition, and rarely exploit powerful modeling
frameworks, like deep neural networks, to capture a variety of complicated
interactions in data. More important, for highly expressive, deep
factorization, we lack an effective approach to handle streaming data, which
are ubiquitous in real-world applications. To address these issues, we propose
SPIDER, a Streaming ProbabilistIc Deep tEnsoR factorization method. We first
use Bayesian neural networks (NNs) to construct a deep tensor factorization
model. We assign a spike-and-slab prior over the NN weights to encourage
sparsity and prevent overfitting. We then use Taylor expansions and moment
matching to approximate the posterior of the NN output and calculate the
running model evidence, based on which we develop an efficient streaming
posterior inference algorithm in the assumed-density-filtering and expectation
propagation framework. Our algorithm provides responsive incremental updates
for the posterior of the latent factors and NN weights upon receiving new
tensor entries, and meanwhile select and inhibit redundant/useless weights. We
show the advantages of our approach in four real-world applications.
Related papers
- Factor Augmented Tensor-on-Tensor Neural Networks [3.0040661953201475]
We propose a Factor Augmented-on-Tensor Neural Network (FATTNN) that integrates tensor factor models into deep neural networks.
We show that our proposed algorithms achieve substantial increases in prediction accuracy and significant reductions in computational time.
arXiv Detail & Related papers (2024-05-30T01:56:49Z) - Regularized PolyKervNets: Optimizing Expressiveness and Efficiency for
Private Inference in Deep Neural Networks [0.0]
We focus on PolyKervNets, a technique known for offering improved dynamic approximations in smaller networks.
Our primary objective is to empirically explore optimization-based training recipes to enhance the performance of PolyKervNets in larger networks.
arXiv Detail & Related papers (2023-12-23T11:37:18Z) - Negative Flux Aggregation to Estimate Feature Attributions [15.411534490483495]
There are increasing demands for understanding deep neural networks' (DNNs) behavior spurred by growing security and/or transparency concerns.
To enhance the explainability of DNNs, we estimate the input feature's attributions to the prediction task using divergence and flux.
Inspired by the divergence theorem in vector analysis, we develop a novel Negative Flux Aggregation (NeFLAG) formulation and an efficient approximation algorithm to estimate attribution map.
arXiv Detail & Related papers (2023-01-17T16:19:41Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Tensor-based framework for training flexible neural networks [9.176056742068813]
We propose a new learning algorithm which solves a constrained coupled matrix-tensor factorization (CMTF) problem.
The proposed algorithm can handle different bases decomposition.
The goal of this method is to compress large pretrained NN models, by replacing tensorworks, em i.e., one or multiple layers of the original network, by a new flexible layer.
arXiv Detail & Related papers (2021-06-25T10:26:48Z) - Non-Gradient Manifold Neural Network [79.44066256794187]
Deep neural network (DNN) generally takes thousands of iterations to optimize via gradient descent.
We propose a novel manifold neural network based on non-gradient optimization.
arXiv Detail & Related papers (2021-06-15T06:39:13Z) - Online Limited Memory Neural-Linear Bandits with Likelihood Matching [53.18698496031658]
We study neural-linear bandits for solving problems where both exploration and representation learning play an important role.
We propose a likelihood matching algorithm that is resilient to catastrophic forgetting and is completely online.
arXiv Detail & Related papers (2021-02-07T14:19:07Z) - Tensor-Train Networks for Learning Predictive Modeling of
Multidimensional Data [0.0]
A promising strategy is based on tensor networks, which have been very successful in physical and chemical applications.
We show that the weights of a multidimensional regression model can be learned by means of tensor networks with the aim of performing a powerful compact representation.
An algorithm based on alternating least squares has been proposed for approximating the weights in TT-format with a reduction of computational power.
arXiv Detail & Related papers (2021-01-22T16:14:38Z) - Phase Retrieval using Expectation Consistent Signal Recovery Algorithm
based on Hypernetwork [73.94896986868146]
Phase retrieval is an important component in modern computational imaging systems.
Recent advances in deep learning have opened up a new possibility for robust and fast PR.
We develop a novel framework for deep unfolding to overcome the existing limitations.
arXiv Detail & Related papers (2021-01-12T08:36:23Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.