Explicit Solution Equation for Every Combinatorial Problem via Tensor Networks: MeLoCoToN
- URL: http://arxiv.org/abs/2502.05981v1
- Date: Sun, 09 Feb 2025 18:16:53 GMT
- Title: Explicit Solution Equation for Every Combinatorial Problem via Tensor Networks: MeLoCoToN
- Authors: Alejandro Mata Ali,
- Abstract summary: We show that every computation problem has an exact explicit equation that returns its solution.
We present a method to obtain an equation that solves exactly any problem, both inversion, constraint satisfaction and optimization.
- Score: 55.2480439325792
- License:
- Abstract: In this paper we show that every combinatorial problem has an exact explicit equation that returns its solution. We present a method to obtain an equation that solves exactly any combinatorial problem, both inversion, constraint satisfaction and optimization, by obtaining its equivalent tensor network. This formulation only requires a basic knowledge of classical logical operators, at a first year level of any computer science degree. These equations are not necessarily computable in a reasonable time, nor do they allow to surpass the state of the art in computational complexity, but they allow to have a new perspective for the mathematical analysis of these problems. These equations computation can be approximated by different methods such as Matrix Product State compression. We also present the equations for numerous combinatorial problems. This work proves that, if there is a physical system capable of contracting in polynomial time the tensor networks presented, every NP-Hard problem can be solved in polynomial time.
Related papers
- Sum-of-Squares inspired Quantum Metaheuristic for Polynomial Optimization with the Hadamard Test and Approximate Amplitude Constraints [76.53316706600717]
Recently proposed quantum algorithm arXiv:2206.14999 is based on semidefinite programming (SDP)
We generalize the SDP-inspired quantum algorithm to sum-of-squares.
Our results show that our algorithm is suitable for large problems and approximate the best known classicals.
arXiv Detail & Related papers (2024-08-14T19:04:13Z) - Towards Quantum Computational Mechanics [1.530480694206666]
We show how quantum computing can be used to solve representative element volume (RVE) problems in computational homogenisation.
Our quantum RVE solver attains exponential acceleration with respect to classical solvers.
arXiv Detail & Related papers (2023-12-06T12:53:02Z) - Fast quantum algorithm for differential equations [0.5895819801677125]
We present a quantum algorithm with numerical complexity that is polylogarithmic in $N$ but is independent of $kappa$ for a large class of PDEs.
Our algorithm generates a quantum state that enables extracting features of the solution.
arXiv Detail & Related papers (2023-06-20T18:01:07Z) - Symmetric Tensor Networks for Generative Modeling and Constrained
Combinatorial Optimization [72.41480594026815]
Constrained optimization problems abound in industry, from portfolio optimization to logistics.
One of the major roadblocks in solving these problems is the presence of non-trivial hard constraints which limit the valid search space.
In this work, we encode arbitrary integer-valued equality constraints of the form Ax=b, directly into U(1) symmetric networks (TNs) and leverage their applicability as quantum-inspired generative models.
arXiv Detail & Related papers (2022-11-16T18:59:54Z) - Tunable Complexity Benchmarks for Evaluating Physics-Informed Neural
Networks on Coupled Ordinary Differential Equations [64.78260098263489]
In this work, we assess the ability of physics-informed neural networks (PINNs) to solve increasingly-complex coupled ordinary differential equations (ODEs)
We show that PINNs eventually fail to produce correct solutions to these benchmarks as their complexity increases.
We identify several reasons why this may be the case, including insufficient network capacity, poor conditioning of the ODEs, and high local curvature, as measured by the Laplacian of the PINN loss.
arXiv Detail & Related papers (2022-10-14T15:01:32Z) - A quantum-inspired tensor network method for constrained combinatorial
optimization problems [5.904219009974901]
We propose a quantum-inspired tensor-network-based algorithm for general locally constrained optimization problems.
Our algorithm constructs a Hamiltonian for the problem of interest, effectively mapping it to a quantum problem.
Our results show the effectiveness of this construction and potential applications.
arXiv Detail & Related papers (2022-03-29T05:44:07Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Power of human-algorithm collaboration in solving combinatorial
optimization problems [0.0]
We show that a class of optimization problems can be solved efficiently in expectation up to a multiplicative factor $epsilon$ where $epsilon$ is arbitrary constant.
While our proposed methods are merely theoretical, they cast new light on how to approach solving these problems that have been usually considered intractable.
arXiv Detail & Related papers (2021-07-25T11:21:59Z) - Efficient Solution of Boolean Satisfiability Problems with Digital
MemComputing [0.0]
We show that a memory-assisted physical system can efficiently solve the SAT problem in continuous time.
The efficiency of the simulations is related to the collective dynamical properties of the original physical system.
We anticipate our results to broaden research directions in physics-inspired computing paradigms.
arXiv Detail & Related papers (2020-11-12T18:13:46Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.