Tensor-Train Networks for Learning Predictive Modeling of
Multidimensional Data
- URL: http://arxiv.org/abs/2101.09184v1
- Date: Fri, 22 Jan 2021 16:14:38 GMT
- Title: Tensor-Train Networks for Learning Predictive Modeling of
Multidimensional Data
- Authors: M. Nazareth da Costa, R. Attux, A. Cichocki, J. M. T. Romano
- Abstract summary: A promising strategy is based on tensor networks, which have been very successful in physical and chemical applications.
We show that the weights of a multidimensional regression model can be learned by means of tensor networks with the aim of performing a powerful compact representation.
An algorithm based on alternating least squares has been proposed for approximating the weights in TT-format with a reduction of computational power.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Deep neural networks have attracted the attention of the machine learning
community because of their appealing data-driven framework and of their
performance in several pattern recognition tasks. On the other hand, there are
many open theoretical problems regarding the internal operation of the network,
the necessity of certain layers, hyperparameter selection etc. A promising
strategy is based on tensor networks, which have been very successful in
physical and chemical applications. In general, higher-order tensors are
decomposed into sparsely interconnected lower-order tensors. This is a
numerically reliable way to avoid the curse of dimensionality and to provide
highly compressed representation of a data tensor, besides the good numerical
properties that allow to control the desired accuracy of approximation. In
order to compare tensor and neural networks, we first consider the
identification of the classical Multilayer Perceptron using Tensor-Train. A
comparative analysis is also carried out in the context of prediction of the
Mackey-Glass noisy chaotic time series and NASDAQ index. We have shown that the
weights of a multidimensional regression model can be learned by means of
tensor networks with the aim of performing a powerful compact representation
retaining the accuracy of neural networks. Furthermore, an algorithm based on
alternating least squares has been proposed for approximating the weights in
TT-format with a reduction of computational calculus. By means of a direct
expression, we have approximated the core estimation as the conventional
solution for a general regression model, which allows to extend the
applicability of tensor structures to different algorithms.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Factor Augmented Tensor-on-Tensor Neural Networks [3.0040661953201475]
We propose a Factor Augmented-on-Tensor Neural Network (FATTNN) that integrates tensor factor models into deep neural networks.
We show that our proposed algorithms achieve substantial increases in prediction accuracy and significant reductions in computational time.
arXiv Detail & Related papers (2024-05-30T01:56:49Z) - Enhancing lattice kinetic schemes for fluid dynamics with Lattice-Equivariant Neural Networks [79.16635054977068]
We present a new class of equivariant neural networks, dubbed Lattice-Equivariant Neural Networks (LENNs)
Our approach develops within a recently introduced framework aimed at learning neural network-based surrogate models Lattice Boltzmann collision operators.
Our work opens towards practical utilization of machine learning-augmented Lattice Boltzmann CFD in real-world simulations.
arXiv Detail & Related papers (2024-05-22T17:23:15Z) - Graph Neural Networks for Learning Equivariant Representations of Neural Networks [55.04145324152541]
We propose to represent neural networks as computational graphs of parameters.
Our approach enables a single model to encode neural computational graphs with diverse architectures.
We showcase the effectiveness of our method on a wide range of tasks, including classification and editing of implicit neural representations.
arXiv Detail & Related papers (2024-03-18T18:01:01Z) - Assessing Neural Network Representations During Training Using
Noise-Resilient Diffusion Spectral Entropy [55.014926694758195]
Entropy and mutual information in neural networks provide rich information on the learning process.
We leverage data geometry to access the underlying manifold and reliably compute these information-theoretic measures.
We show that they form noise-resistant measures of intrinsic dimensionality and relationship strength in high-dimensional simulated data.
arXiv Detail & Related papers (2023-12-04T01:32:42Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Tensor Decomposition for Model Reduction in Neural Networks: A Review [13.96938227911258]
Modern neural networks have revolutionized the fields of computer vision (CV) and Natural Language Processing (NLP)
They are widely used for solving complex CV tasks and NLP tasks such as image classification, image generation, and machine translation.
This paper reviews six tensor decomposition methods and illustrates their ability to compress model parameters.
arXiv Detail & Related papers (2023-04-26T13:12:00Z) - Tensor Networks for Multi-Modal Non-Euclidean Data [24.50116388903113]
We introduce a novel Multi-Graph Network (MGTN) framework, which leverages on the desirable properties of graphs, tensors and neural networks in a physically meaningful and compact manner.
This equips MGTNs with the ability to exploit local information in irregular data sources at a drastically reduced parameter complexity.
The benefits of the MGTN framework, especially its ability to avoid overfitting through the inherent low-rank regularization properties of tensor networks, are demonstrated.
arXiv Detail & Related papers (2021-03-27T21:33:46Z) - Segmenting two-dimensional structures with strided tensor networks [1.952097552284465]
We propose a novel formulation of tensor networks for supervised image segmentation.
The proposed model is end-to-end trainable using backpropagation.
The evaluation shows that the strided tensor network yields competitive performance compared to CNN-based models.
arXiv Detail & Related papers (2021-02-13T11:06:34Z) - Anomaly Detection with Tensor Networks [2.3895981099137535]
We exploit the memory and computational efficiency of tensor networks to learn a linear transformation over a space with a dimension exponential in the number of original features.
We produce competitive results on image datasets, despite not exploiting the locality of images.
arXiv Detail & Related papers (2020-06-03T20:41:30Z) - Understanding Generalization in Deep Learning via Tensor Methods [53.808840694241]
We advance the understanding of the relations between the network's architecture and its generalizability from the compression perspective.
We propose a series of intuitive, data-dependent and easily-measurable properties that tightly characterize the compressibility and generalizability of neural networks.
arXiv Detail & Related papers (2020-01-14T22:26:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.