Neural approximation of Wasserstein distance via a universal
architecture for symmetric and factorwise group invariant functions
- URL: http://arxiv.org/abs/2308.00273v2
- Date: Fri, 17 Nov 2023 08:37:18 GMT
- Title: Neural approximation of Wasserstein distance via a universal
architecture for symmetric and factorwise group invariant functions
- Authors: Samantha Chen, Yusu Wang
- Abstract summary: We first present a general neural network architecture for approximating SFGI functions.
The main contribution of this paper combines this general neural network with a sketching idea to develop a specific and efficient neural network.
Our work provides an interesting integration of sketching ideas for geometric problems with universal approximation of symmetric functions.
- Score: 6.994580267603235
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning distance functions between complex objects, such as the Wasserstein
distance to compare point sets, is a common goal in machine learning
applications. However, functions on such complex objects (e.g., point sets and
graphs) are often required to be invariant to a wide variety of group actions
e.g. permutation or rigid transformation. Therefore, continuous and symmetric
product functions (such as distance functions) on such complex objects must
also be invariant to the product of such group actions. We call these functions
symmetric and factor-wise group invariant (or SFGI functions in short). In this
paper, we first present a general neural network architecture for approximating
SFGI functions. The main contribution of this paper combines this general
neural network with a sketching idea to develop a specific and efficient neural
network which can approximate the $p$-th Wasserstein distance between point
sets. Very importantly, the required model complexity is independent of the
sizes of input point sets. On the theoretical front, to the best of our
knowledge, this is the first result showing that there exists a neural network
with the capacity to approximate Wasserstein distance with bounded model
complexity. Our work provides an interesting integration of sketching ideas for
geometric problems with universal approximation of symmetric functions. On the
empirical front, we present a range of results showing that our newly proposed
neural network architecture performs comparatively or better than other models
(including a SOTA Siamese Autoencoder based approach). In particular, our
neural network generalizes significantly better and trains much faster than the
SOTA Siamese AE. Finally, this line of investigation could be useful in
exploring effective neural network design for solving a broad range of
geometric optimization problems (e.g., $k$-means in a metric space).
Related papers
- Towards Explaining Hypercomplex Neural Networks [6.543091030789653]
Hypercomplex neural networks are gaining increasing interest in the deep learning community.
In this paper, we propose inherently interpretable PHNNs and quaternion-like networks.
We draw insights into how this unique branch of neural models operates.
arXiv Detail & Related papers (2024-03-26T17:58:07Z) - Universal Neural Functionals [67.80283995795985]
A challenging problem in many modern machine learning tasks is to process weight-space features.
Recent works have developed promising weight-space models that are equivariant to the permutation symmetries of simple feedforward networks.
This work proposes an algorithm that automatically constructs permutation equivariant models for any weight space.
arXiv Detail & Related papers (2024-02-07T20:12:27Z) - Permutation Equivariant Neural Functionals [92.0667671999604]
This work studies the design of neural networks that can process the weights or gradients of other neural networks.
We focus on the permutation symmetries that arise in the weights of deep feedforward networks because hidden layer neurons have no inherent order.
In our experiments, we find that permutation equivariant neural functionals are effective on a diverse set of tasks.
arXiv Detail & Related papers (2023-02-27T18:52:38Z) - Efficient Parametric Approximations of Neural Network Function Space
Distance [6.117371161379209]
It is often useful to compactly summarize important properties of model parameters and training data so that they can be used later without storing and/or iterating over the entire dataset.
We consider estimating the Function Space Distance (FSD) over a training set, i.e. the average discrepancy between the outputs of two neural networks.
We propose a Linearized Activation TRick (LAFTR) and derive an efficient approximation to FSD for ReLU neural networks.
arXiv Detail & Related papers (2023-02-07T15:09:23Z) - Dist2Cycle: A Simplicial Neural Network for Homology Localization [66.15805004725809]
Simplicial complexes can be viewed as high dimensional generalizations of graphs that explicitly encode multi-way ordered relations.
We propose a graph convolutional model for learning functions parametrized by the $k$-homological features of simplicial complexes.
arXiv Detail & Related papers (2021-10-28T14:59:41Z) - Frame Averaging for Invariant and Equivariant Network Design [50.87023773850824]
We introduce Frame Averaging (FA), a framework for adapting known (backbone) architectures to become invariant or equivariant to new symmetry types.
We show that FA-based models have maximal expressive power in a broad setting.
We propose a new class of universal Graph Neural Networks (GNNs), universal Euclidean motion invariant point cloud networks, and Euclidean motion invariant Message Passing (MP) GNNs.
arXiv Detail & Related papers (2021-10-07T11:05:23Z) - Statistically Meaningful Approximation: a Case Study on Approximating
Turing Machines with Transformers [50.85524803885483]
This work proposes a formal definition of statistically meaningful (SM) approximation which requires the approximating network to exhibit good statistical learnability.
We study SM approximation for two function classes: circuits and Turing machines.
arXiv Detail & Related papers (2021-07-28T04:28:55Z) - A Functional Perspective on Learning Symmetric Functions with Neural
Networks [48.80300074254758]
We study the learning and representation of neural networks defined on measures.
We establish approximation and generalization bounds under different choices of regularization.
The resulting models can be learned efficiently and enjoy generalization guarantees that extend across input sizes.
arXiv Detail & Related papers (2020-08-16T16:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.