Tensor networks for unsupervised machine learning
- URL: http://arxiv.org/abs/2106.12974v1
- Date: Thu, 24 Jun 2021 12:51:00 GMT
- Title: Tensor networks for unsupervised machine learning
- Authors: Jing Liu, Sujie Li, Jiang Zhang, Pan Zhang
- Abstract summary: We present the Autoregressive Matrix Product States (AMPS), a tensor-network-based model combining the matrix product states from quantum many-body physics and the autoregressive models from machine learning.
We show that the proposed model significantly outperforms the existing tensor-network-based models and the restricted Boltzmann machines.
- Score: 9.897828174118974
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modeling the joint distribution of high-dimensional data is a central task in
unsupervised machine learning. In recent years, many interests have been
attracted to developing learning models based on tensor networks, which have
advantages of theoretical understandings of the expressive power using
entanglement properties, and as a bridge connecting the classical computation
and the quantum computation. Despite the great potential, however, existing
tensor-network-based unsupervised models only work as a proof of principle, as
their performances are much worse than the standard models such as the
restricted Boltzmann machines and neural networks. In this work, we present the
Autoregressive Matrix Product States (AMPS), a tensor-network-based model
combining the matrix product states from quantum many-body physics and the
autoregressive models from machine learning. The model enjoys exact calculation
of normalized probability and unbiased sampling, as well as a clear theoretical
understanding of expressive power. We demonstrate the performance of our model
using two applications, the generative modeling on synthetic and real-world
data, and the reinforcement learning in statistical physics. Using extensive
numerical experiments, we show that the proposed model significantly
outperforms the existing tensor-network-based models and the restricted
Boltzmann machines, and is competitive with the state-of-the-art neural network
models.
Related papers
- Generative Learning of Continuous Data by Tensor Networks [45.49160369119449]
We introduce a new family of tensor network generative models for continuous data.
We benchmark the performance of this model on several synthetic and real-world datasets.
Our methods give important theoretical and empirical evidence of the efficacy of quantum-inspired methods for the rapidly growing field of generative learning.
arXiv Detail & Related papers (2023-10-31T14:37:37Z) - The cross-sectional stock return predictions via quantum neural network
and tensor network [0.0]
We investigate the application of quantum and quantum-inspired machine learning algorithms to stock return predictions.
We evaluate the performance of quantum neural network, an algorithm suited for noisy intermediate-scale quantum computers, and tensor network, a quantum-inspired machine learning algorithm.
arXiv Detail & Related papers (2023-04-25T00:05:13Z) - Vertical Layering of Quantized Neural Networks for Heterogeneous
Inference [57.42762335081385]
We study a new vertical-layered representation of neural network weights for encapsulating all quantized models into a single one.
We can theoretically achieve any precision network for on-demand service while only needing to train and maintain one model.
arXiv Detail & Related papers (2022-12-10T15:57:38Z) - Real-time Neural-MPC: Deep Learning Model Predictive Control for
Quadrotors and Agile Robotic Platforms [59.03426963238452]
We present Real-time Neural MPC, a framework to efficiently integrate large, complex neural network architectures as dynamics models within a model-predictive control pipeline.
We show the feasibility of our framework on real-world problems by reducing the positional tracking error by up to 82% when compared to state-of-the-art MPC approaches without neural network dynamics.
arXiv Detail & Related papers (2022-03-15T09:38:15Z) - EINNs: Epidemiologically-Informed Neural Networks [75.34199997857341]
We introduce a new class of physics-informed neural networks-EINN-crafted for epidemic forecasting.
We investigate how to leverage both the theoretical flexibility provided by mechanistic models as well as the data-driven expressability afforded by AI models.
arXiv Detail & Related papers (2022-02-21T18:59:03Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - On Energy-Based Models with Overparametrized Shallow Neural Networks [44.74000986284978]
Energy-based models (EBMs) are a powerful framework for generative modeling.
In this work we focus on shallow neural networks.
We show that models trained in the so-called "active" regime provide a statistical advantage over their associated "lazy" or kernel regime.
arXiv Detail & Related papers (2021-04-15T15:34:58Z) - From Boltzmann Machines to Neural Networks and Back Again [31.613544605376624]
We give new results for learning Restricted Boltzmann Machines, probably the most well-studied class of latent variable models.
Our results are based on new connections to learning two-layer neural networks under $ell_infty$ bounded input.
We then give an algorithm for learning a natural class of supervised RBMs with better runtime than what is possible for its related class of networks without distributional assumptions.
arXiv Detail & Related papers (2020-07-25T00:42:50Z) - Learning Queuing Networks by Recurrent Neural Networks [0.0]
We propose a machine-learning approach to derive performance models from data.
We exploit a deterministic approximation of their average dynamics in terms of a compact system of ordinary differential equations.
This allows for an interpretable structure of the neural network, which can be trained from system measurements to yield a white-box parameterized model.
arXiv Detail & Related papers (2020-02-25T10:56:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.