Connecting Dualities and Machine Learning
- URL: http://arxiv.org/abs/2002.05169v1
- Date: Wed, 12 Feb 2020 19:00:02 GMT
- Title: Connecting Dualities and Machine Learning
- Authors: Philip Betzler, Sven Krippendorf
- Abstract summary: Dualities are widely used in quantum field theories and string theory to obtain correlation functions at high accuracy.
Here we present examples where dual data representations are useful in supervised classification, linking machine learning and typical tasks in theoretical physics.
We find that additional contributions to the loss based on feature separation, feature matching with respect to desired representations, and a good performance on a simple' correlation function can lead to known and unknown dual representations.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dualities are widely used in quantum field theories and string theory to
obtain correlation functions at high accuracy. Here we present examples where
dual data representations are useful in supervised classification, linking
machine learning and typical tasks in theoretical physics. We then discuss how
such beneficial representations can be enforced in the latent dimension of
neural networks. We find that additional contributions to the loss based on
feature separation, feature matching with respect to desired representations,
and a good performance on a `simple' correlation function can lead to known and
unknown dual representations. This is the first proof of concept that computers
can find dualities. We discuss how our examples, based on discrete Fourier
transformation and Ising models, connect to other dualities in theoretical
physics, for instance Seiberg duality.
Related papers
- Machine learning and optimization-based approaches to duality in statistical physics [2.3727769223905515]
duality is the idea that a given physical system can have two different mathematical descriptions.
We numerically solve the problem and show that our framework can rediscover the celebrated Kramers-Wannier duality for the 2d Ising model.
We also discuss an alternative approach which uses known features of the mapping of topological lines to reduce the problem to optimize the couplings in a dual Hamiltonian.
arXiv Detail & Related papers (2024-11-07T16:29:03Z) - Bridging Associative Memory and Probabilistic Modeling [29.605203018237457]
Associative memory and probabilistic modeling are two fundamental topics in artificial intelligence.
We build a bridge between the two that enables useful flow of ideas in both directions.
arXiv Detail & Related papers (2024-02-15T18:56:46Z) - Going Beyond Neural Network Feature Similarity: The Network Feature
Complexity and Its Interpretation Using Category Theory [64.06519549649495]
We provide the definition of what we call functionally equivalent features.
These features produce equivalent output under certain transformations.
We propose an efficient algorithm named Iterative Feature Merging.
arXiv Detail & Related papers (2023-10-10T16:27:12Z) - Topological dualities via tensor networks [0.0]
Ground state of the toric code, that of the two-dimensional class D superconductor, and the partition sum of the two-dimensional Ising model are dual to each other.
Connecting fermionic and bosonic systems, the duality construction is intrinsically non-local.
We propose a unified approach to this duality, whose main protagonist is a tensor network (TN) assuming the role of an intermediate translator.
arXiv Detail & Related papers (2023-09-22T18:00:17Z) - A simple probabilistic neural network for machine understanding [0.0]
We discuss probabilistic neural networks with a fixed internal representation as models for machine understanding.
We derive the internal representation by requiring that it satisfies the principles of maximal relevance and of maximal ignorance about how different features are combined.
We argue that learning machines with this architecture enjoy a number of interesting properties, like the continuity of the representation with respect to changes in parameters and data.
arXiv Detail & Related papers (2022-10-24T13:00:15Z) - Curvature-informed multi-task learning for graph networks [56.155331323304]
State-of-the-art graph neural networks attempt to predict multiple properties simultaneously.
We investigate a potential explanation for this phenomenon: the curvature of each property's loss surface significantly varies, leading to inefficient learning.
arXiv Detail & Related papers (2022-08-02T18:18:41Z) - On Neural Architecture Inductive Biases for Relational Tasks [76.18938462270503]
We introduce a simple architecture based on similarity-distribution scores which we name Compositional Network generalization (CoRelNet)
We find that simple architectural choices can outperform existing models in out-of-distribution generalizations.
arXiv Detail & Related papers (2022-06-09T16:24:01Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Can Temporal-Difference and Q-Learning Learn Representation? A Mean-Field Theory [110.99247009159726]
Temporal-difference and Q-learning play a key role in deep reinforcement learning, where they are empowered by expressive nonlinear function approximators such as neural networks.
In particular, temporal-difference learning converges when the function approximator is linear in a feature representation, which is fixed throughout learning, and possibly diverges otherwise.
arXiv Detail & Related papers (2020-06-08T17:25:22Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.