Operationally meaningful representations of physical systems in neural
networks
- URL: http://arxiv.org/abs/2001.00593v1
- Date: Thu, 2 Jan 2020 19:01:31 GMT
- Title: Operationally meaningful representations of physical systems in neural
networks
- Authors: Hendrik Poulsen Nautrup, Tony Metger, Raban Iten, Sofiene Jerbi, Lea
M. Trenkwalder, Henrik Wilming, Hans J. Briegel, Renato Renner
- Abstract summary: We present a neural network architecture based on the notion that agents dealing with different aspects of a physical system should be able to communicate relevant information as efficiently as possible to one another.
This produces representations that separate different parameters which are useful for making statements about the physical system in different experimental settings.
- Score: 4.192302677744796
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To make progress in science, we often build abstract representations of
physical systems that meaningfully encode information about the systems. The
representations learnt by most current machine learning techniques reflect
statistical structure present in the training data; however, these methods do
not allow us to specify explicit and operationally meaningful requirements on
the representation. Here, we present a neural network architecture based on the
notion that agents dealing with different aspects of a physical system should
be able to communicate relevant information as efficiently as possible to one
another. This produces representations that separate different parameters which
are useful for making statements about the physical system in different
experimental settings. We present examples involving both classical and quantum
physics. For instance, our architecture finds a compact representation of an
arbitrary two-qubit system that separates local parameters from parameters
describing quantum correlations. We further show that this method can be
combined with reinforcement learning to enable representation learning within
interactive scenarios where agents need to explore experimental settings to
identify relevant variables.
Related papers
- Interpretable Meta-Learning of Physical Systems [4.343110120255532]
Recent meta-learning methods rely on black-box neural networks, resulting in high computational costs and limited interpretability.
We argue that multi-environment generalization can be achieved using a simpler learning model, with an affine structure with respect to the learning task.
We demonstrate the competitive generalization performance and the low computational cost of our method by comparing it to state-of-the-art algorithms on physical systems.
arXiv Detail & Related papers (2023-12-01T10:18:50Z) - Do Neural Networks Trained with Topological Features Learn Different
Internal Representations? [1.418465438044804]
We investigate whether a model trained with topological features learns internal representations of data that are fundamentally different than those learned by a model trained with the original raw data.
We find that structurally, the hidden representations of models trained and evaluated on topological features differ substantially compared to those trained and evaluated on the corresponding raw data.
We conjecture that this means that neural networks trained on raw data may extract some limited topological features in the process of making predictions.
arXiv Detail & Related papers (2022-11-14T19:19:04Z) - Neural Implicit Representations for Physical Parameter Inference from a Single Video [49.766574469284485]
We propose to combine neural implicit representations for appearance modeling with neural ordinary differential equations (ODEs) for modelling physical phenomena.
Our proposed model combines several unique advantages: (i) Contrary to existing approaches that require large training datasets, we are able to identify physical parameters from only a single video.
The use of neural implicit representations enables the processing of high-resolution videos and the synthesis of photo-realistic images.
arXiv Detail & Related papers (2022-04-29T11:55:35Z) - Physical Modeling using Recurrent Neural Networks with Fast
Convolutional Layers [1.7013938542585922]
We describe several novel recurrent neural network structures and show how they can be thought of as an extension of modal techniques.
As a proof of concept, we generate synthetic data for three physical systems and show that the proposed network structures can be trained with this data to reproduce the behavior of these systems.
arXiv Detail & Related papers (2022-04-21T14:22:44Z) - Learning Dynamics and Structure of Complex Systems Using Graph Neural
Networks [13.509027957413409]
We trained graph neural networks to fit time series from an example nonlinear dynamical system.
We found simple interpretations of the learned representation and model components.
We successfully identified a graph translator' between the statistical interactions in belief propagation and parameters of the corresponding trained network.
arXiv Detail & Related papers (2022-02-22T15:58:16Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Dynamic Inference with Neural Interpreters [72.90231306252007]
We present Neural Interpreters, an architecture that factorizes inference in a self-attention network as a system of modules.
inputs to the model are routed through a sequence of functions in a way that is end-to-end learned.
We show that Neural Interpreters perform on par with the vision transformer using fewer parameters, while being transferrable to a new task in a sample efficient manner.
arXiv Detail & Related papers (2021-10-12T23:22:45Z) - Discrete-Valued Neural Communication [85.3675647398994]
We show that restricting the transmitted information among components to discrete representations is a beneficial bottleneck.
Even though individuals have different understandings of what a "cat" is based on their specific experiences, the shared discrete token makes it possible for communication among individuals to be unimpeded by individual differences in internal representation.
We extend the quantization mechanism from the Vector-Quantized Variational Autoencoder to multi-headed discretization with shared codebooks and use it for discrete-valued neural communication.
arXiv Detail & Related papers (2021-07-06T03:09:25Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - A Framework for Learning Invariant Physical Relations in Multimodal
Sensory Processing [0.0]
We design a novel neural network architecture capable of learning, in an unsupervised manner, relations among sensory cues.
We describe the core system functionality when learning arbitrary non-linear relations in low-dimensional sensory data.
We demonstrate this through a real-world learning problem, where, from standard RGB camera frames, the network learns the relations between physical quantities.
arXiv Detail & Related papers (2020-06-30T08:42:48Z) - End-to-End Models for the Analysis of System 1 and System 2 Interactions
based on Eye-Tracking Data [99.00520068425759]
We propose a computational method, within a modified visual version of the well-known Stroop test, for the identification of different tasks and potential conflicts events.
A statistical analysis shows that the selected variables can characterize the variation of attentive load within different scenarios.
We show that Machine Learning techniques allow to distinguish between different tasks with a good classification accuracy.
arXiv Detail & Related papers (2020-02-03T17:46:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.