Doubly Reparameterized Importance Weighted Structure Learning for Scene
Graph Generation
- URL: http://arxiv.org/abs/2206.11352v1
- Date: Wed, 22 Jun 2022 20:00:25 GMT
- Title: Doubly Reparameterized Importance Weighted Structure Learning for Scene
Graph Generation
- Authors: Daqi Liu, Miroslaw Bober, Josef Kittler
- Abstract summary: Scene graph generation, given an input image, aims to explicitly model objects and their relationships by constructing a visually-grounded scene graph.
We propose a novel doubly re parameterized importance weighted structure learning method, which employs a tighter importance weighted lower bound as the variational inference objective.
The proposed method achieves the state-of-the-art performance on various popular scene graph generation benchmarks.
- Score: 40.46394569128303
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a structured prediction task, scene graph generation, given an input
image, aims to explicitly model objects and their relationships by constructing
a visually-grounded scene graph. In the current literature, such task is
universally solved via a message passing neural network based mean field
variational Bayesian methodology. The classical loose evidence lower bound is
generally chosen as the variational inference objective, which could induce
oversimplified variational approximation and thus underestimate the underlying
complex posterior. In this paper, we propose a novel doubly reparameterized
importance weighted structure learning method, which employs a tighter
importance weighted lower bound as the variational inference objective. It is
computed from multiple samples drawn from a reparameterizable Gumbel-Softmax
sampler and the resulting constrained variational inference task is solved by a
generic entropic mirror descent algorithm. The resulting doubly reparameterized
gradient estimator reduces the variance of the corresponding derivatives with a
beneficial impact on learning. The proposed method achieves the
state-of-the-art performance on various popular scene graph generation
benchmarks.
Related papers
- Joint Graph Learning and Model Fitting in Laplacian Regularized
Stratified Models [5.933030735757292]
Laplacian regularized stratified models (LRSM) are models that utilize the explicit or implicit network structure of the sub-problems.
This paper shows the importance and sensitivity of graph weights in LRSM, and provably show that the sensitivity can be arbitrarily large.
We propose a generic approach to jointly learn the graph while fitting the model parameters by solving a single optimization problem.
arXiv Detail & Related papers (2023-05-04T06:06:29Z) - Importance Weighted Structure Learning for Scene Graph Generation [40.46394569128303]
We propose a novel importance weighted structure learning method for scene graph generation.
A generic entropic mirror descent algorithm is applied to solve the resulting constrained variational inference task.
The proposed method achieves the state-of-the-art performance on various popular scene graph generation benchmarks.
arXiv Detail & Related papers (2022-05-14T09:25:14Z) - Constrained Structure Learning for Scene Graph Generation [40.46394569128303]
We present a constrained structure learning method, for which an explicit constrained variational inference objective is proposed.
We validate the proposed generic model on various popular scene graph generation benchmarks and show that it outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2022-01-27T17:47:37Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - On the Role of Optimization in Double Descent: A Least Squares Study [30.44215064390409]
We show an excess risk bound for the descent gradient solution of the least squares objective.
We find that in case of noiseless regression, double descent is explained solely by optimization-related quantities.
We empirically explore if our predictions hold for neural networks.
arXiv Detail & Related papers (2021-07-27T09:13:11Z) - Instance-Level Relative Saliency Ranking with Graph Reasoning [126.09138829920627]
We present a novel unified model to segment salient instances and infer relative saliency rank order.
A novel loss function is also proposed to effectively train the saliency ranking branch.
experimental results demonstrate that our proposed model is more effective than previous methods.
arXiv Detail & Related papers (2021-07-08T13:10:42Z) - Deep Magnification-Flexible Upsampling over 3D Point Clouds [103.09504572409449]
We propose a novel end-to-end learning-based framework to generate dense point clouds.
We first formulate the problem explicitly, which boils down to determining the weights and high-order approximation errors.
Then, we design a lightweight neural network to adaptively learn unified and sorted weights as well as the high-order refinements.
arXiv Detail & Related papers (2020-11-25T14:00:18Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.