A Fast Parallel Tensor Decomposition with Optimal Stochastic Gradient
Descent: an Application in Structural Damage Identification
- URL: http://arxiv.org/abs/2111.02632v1
- Date: Thu, 4 Nov 2021 05:17:07 GMT
- Title: A Fast Parallel Tensor Decomposition with Optimal Stochastic Gradient
Descent: an Application in Structural Damage Identification
- Authors: Ali Anaissi, Basem Suleiman and Seid Miad Zandavi
- Abstract summary: We propose a novel algorithm, FP-CPD, to parallelize the CANDECOMP/PARAFAC (CP) decomposition of a tensor $mathcalX in mathbbR I_1 times dots times I_N $.
- Score: 1.536989504296526
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Structural Health Monitoring (SHM) provides an economic approach which aims
to enhance understanding the behavior of structures by continuously collects
data through multiple networked sensors attached to the structure. This data is
then utilized to gain insight into the health of a structure and make timely
and economic decisions about its maintenance. The generated SHM sensing data is
non-stationary and exists in a correlated multi-way form which makes the
batch/off-line learning and standard two-way matrix analysis unable to capture
all of these correlations and relationships. In this sense, the online tensor
data analysis has become an essential tool for capturing underlying structures
in higher-order datasets stored in a tensor $\mathcal{X} \in \mathbb{R} ^{I_1
\times \dots \times I_N} $. The CANDECOMP/PARAFAC (CP) decomposition has been
extensively studied and applied to approximate X by N loading matrices A(1), .
. . ,A(N) where N represents the order of the tensor. We propose a novel
algorithm, FP-CPD, to parallelize the CANDECOMP/PARAFAC (CP) decomposition of a
tensor $\mathcal{X} \in \mathbb{R} ^{I_1 \times \dots \times I_N} $. Our
approach is based on stochastic gradient descent (SGD) algorithm which allows
us to parallelize the learning process and it is very useful in online setting
since it updates $\mathcal{X}^{t+1}$ in one single step. Our SGD algorithm is
augmented with Nesterov's Accelerated Gradient (NAG) and perturbation methods
to accelerate and guarantee convergence. The experimental results using
laboratory-based and real-life structural datasets indicate fast convergence
and good scalability.
Related papers
- Explainable Spatio-Temporal GCNNs for Irregular Multivariate Time Series: Architecture and Application to ICU Patient Data [7.433698348783128]
We present XST-CNN (eXG-Temporal Graph Conal Neural Network), a novel architecture for processing heterogeneous and irregular Multi Time Series (MTS) data.
Our approach captures temporal and feature within a unifiedtemporal-temporal pipeline by leveraging a GCNN pipeline.
We evaluate XST-CNN using real-world Electronic Health Record data to predict Multidrug Resistance (MDR) in ICU patients.
arXiv Detail & Related papers (2024-11-01T22:53:17Z) - Compressing Structured Tensor Algebra [1.2624532490634646]
DASTAC is a framework to propagate the tensors's captured high-level structure down to low-level code generation.
Our methodology reduces memory footprint by automatically detecting the best data layout.
We show that DASTAC achieves 1 to 2 orders of magnitude speedup over TACO, a state-of-the-art sparse tensor compiler, and StructTensor, a state-of-the-art structured tensor algebra compiler.
arXiv Detail & Related papers (2024-07-18T17:25:17Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Stochastic Optimization for Non-convex Problem with Inexact Hessian
Matrix, Gradient, and Function [99.31457740916815]
Trust-region (TR) and adaptive regularization using cubics have proven to have some very appealing theoretical properties.
We show that TR and ARC methods can simultaneously provide inexact computations of the Hessian, gradient, and function values.
arXiv Detail & Related papers (2023-10-18T10:29:58Z) - Fast and Provable Tensor Robust Principal Component Analysis via Scaled
Gradient Descent [30.299284742925852]
This paper tackles tensor robust principal component analysis (RPCA)
It aims to recover a low-rank tensor from its observations contaminated by sparse corruptions.
We show that the proposed algorithm achieves better and more scalable performance than state-of-the-art matrix and tensor RPCA algorithms.
arXiv Detail & Related papers (2022-06-18T04:01:32Z) - Truncated tensor Schatten p-norm based approach for spatiotemporal
traffic data imputation with complicated missing patterns [77.34726150561087]
We introduce four complicated missing patterns, including missing and three fiber-like missing cases according to the mode-drivenn fibers.
Despite nonity of the objective function in our model, we derive the optimal solutions by integrating alternating data-mputation method of multipliers.
arXiv Detail & Related papers (2022-05-19T08:37:56Z) - Multi-task Learning of Order-Consistent Causal Graphs [59.9575145128345]
We consider the problem of discovering $K related Gaussian acyclic graphs (DAGs)
Under multi-task learning setting, we propose a $l_1/l$-regularized maximum likelihood estimator (MLE) for learning $K$ linear structural equation models.
We theoretically show that the joint estimator, by leveraging data across related tasks, can achieve a better sample complexity for recovering the causal order.
arXiv Detail & Related papers (2021-11-03T22:10:18Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - Online nonnegative CP-dictionary learning for Markovian data [8.490619842547739]
We introduce a novel algorithm that learns a CANDECOMP/PARAFAC basis from a given stream of tensor-valued data under general constraints.
We prove that our algorithm converges almost surely to the set of stationary points of the objective function under the hypothesis that the sequence of data tensors is generated by an underlying Markov chain.
arXiv Detail & Related papers (2020-09-16T11:41:01Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - NeCPD: An Online Tensor Decomposition with Optimal Stochastic Gradient
Descent [1.0953917735844645]
We propose a new efficient decomposition algorithm named NeCPD for non- efficient problem in multi-way online data based $(N)N.
We further apply this method in the field of reallife monitoring using structural datasets.
arXiv Detail & Related papers (2020-03-18T04:44:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.