Graphical Normalizing Flows
- URL: http://arxiv.org/abs/2006.02548v3
- Date: Fri, 12 Feb 2021 16:47:28 GMT
- Title: Graphical Normalizing Flows
- Authors: Antoine Wehenkel and Gilles Louppe
- Abstract summary: Normalizing flows model complex probability distributions by combining a base distribution with a series of neural networks.
State-of-the-art architectures rely on coupling and autoregressive transformations to lift up invertible functions from scalars to vectors.
We propose the graphical normalizing flow, a new invertible transformation with either a prescribed or a learnable graphical structure.
- Score: 11.23030807455021
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Normalizing flows model complex probability distributions by combining a base
distribution with a series of bijective neural networks. State-of-the-art
architectures rely on coupling and autoregressive transformations to lift up
invertible functions from scalars to vectors. In this work, we revisit these
transformations as probabilistic graphical models, showing they reduce to
Bayesian networks with a pre-defined topology and a learnable density at each
node. From this new perspective, we propose the graphical normalizing flow, a
new invertible transformation with either a prescribed or a learnable graphical
structure. This model provides a promising way to inject domain knowledge into
normalizing flows while preserving both the interpretability of Bayesian
networks and the representation capacity of normalizing flows. We show that
graphical conditioners discover relevant graph structure when we cannot
hypothesize it. In addition, we analyze the effect of $\ell_1$-penalization on
the recovered structure and on the quality of the resulting density estimation.
Finally, we show that graphical conditioners lead to competitive white box
density estimators. Our implementation is available at
https://github.com/AWehenkel/DAG-NF.
Related papers
- Interpretable A-posteriori Error Indication for Graph Neural Network Surrogate Models [0.0]
This work introduces an interpretability enhancement procedure for graph neural networks (GNNs)
The end result is an interpretable GNN model that isolates regions in physical space, corresponding to sub-graphs, that are intrinsically linked to the forecasting task.
The interpretable GNNs can also be used to identify, during inference, graph nodes that correspond to a majority of the anticipated forecasting error.
arXiv Detail & Related papers (2023-11-13T18:37:07Z) - Advective Diffusion Transformers for Topological Generalization in Graph
Learning [69.2894350228753]
We show how graph diffusion equations extrapolate and generalize in the presence of varying graph topologies.
We propose a novel graph encoder backbone, Advective Diffusion Transformer (ADiT), inspired by advective graph diffusion equations.
arXiv Detail & Related papers (2023-10-10T08:40:47Z) - Geometric Graph Filters and Neural Networks: Limit Properties and
Discriminability Trade-offs [122.06927400759021]
We study the relationship between a graph neural network (GNN) and a manifold neural network (MNN) when the graph is constructed from a set of points sampled from the manifold.
We prove non-asymptotic error bounds showing that convolutional filters and neural networks on these graphs converge to convolutional filters and neural networks on the continuous manifold.
arXiv Detail & Related papers (2023-05-29T08:27:17Z) - GrannGAN: Graph annotation generative adversarial networks [72.66289932625742]
We consider the problem of modelling high-dimensional distributions and generating new examples of data with complex relational feature structure coherent with a graph skeleton.
The model we propose tackles the problem of generating the data features constrained by the specific graph structure of each data point by splitting the task into two phases.
In the first it models the distribution of features associated with the nodes of the given graph, in the second it complements the edge features conditionally on the node features.
arXiv Detail & Related papers (2022-12-01T11:49:07Z) - Capturing Graphs with Hypo-Elliptic Diffusions [7.704064306361941]
We show that the distribution of random walks evolves according to a diffusion equation defined using the graph Laplacian.
This results in a novel tensor-valued graph operator, which we call the hypo-elliptic graph Laplacian.
We show that this method competes with graph transformers on datasets requiring long-range reasoning but scales only linearly in the number of edges.
arXiv Detail & Related papers (2022-05-27T16:47:34Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - You say Normalizing Flows I see Bayesian Networks [11.23030807455021]
We show that normalizing flows reduce to Bayesian networks with a pre-defined topology and a learnable density at each node.
We show that stacking multiple transformations in a normalizing flow relaxes independence assumptions and entangles the model distribution.
We prove the non-universality of the affine normalizing flow, regardless of its depth.
arXiv Detail & Related papers (2020-06-01T11:54:50Z) - Residual Correlation in Graph Neural Network Regression [39.54530450932135]
We show that conditional independence assumption severely limits predictive power.
We address this problem with an interpretable and efficient framework.
Our framework achieves substantially higher accuracy than competing baselines.
arXiv Detail & Related papers (2020-02-19T16:32:54Z) - Semi-Supervised Learning with Normalizing Flows [54.376602201489995]
FlowGMM is an end-to-end approach to generative semi supervised learning with normalizing flows.
We show promising results on a wide range of applications, including AG-News and Yahoo Answers text data.
arXiv Detail & Related papers (2019-12-30T17:36:33Z) - Understanding Graph Neural Networks with Generalized Geometric
Scattering Transforms [67.88675386638043]
The scattering transform is a multilayered wavelet-based deep learning architecture that acts as a model of convolutional neural networks.
We introduce windowed and non-windowed geometric scattering transforms for graphs based upon a very general class of asymmetric wavelets.
We show that these asymmetric graph scattering transforms have many of the same theoretical guarantees as their symmetric counterparts.
arXiv Detail & Related papers (2019-11-14T17:23:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.