Solving Visual Analogies Using Neural Algorithmic Reasoning
- URL: http://arxiv.org/abs/2111.10361v1
- Date: Fri, 19 Nov 2021 18:48:16 GMT
- Title: Solving Visual Analogies Using Neural Algorithmic Reasoning
- Authors: Atharv Sonwane, Gautam Shroff, Lovekesh Vig, Ashwin Srinivasan,
Tirtharaj Dash
- Abstract summary: We search for a sequence of elementary neural network transformations that manipulate distributed representations derived from a symbolic space.
We evaluate the extent to which our neural reasoning' approach generalizes for images with unseen shapes and positions.
- Score: 22.384921045720752
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider a class of visual analogical reasoning problems that involve
discovering the sequence of transformations by which pairs of input/output
images are related, so as to analogously transform future inputs. This program
synthesis task can be easily solved via symbolic search. Using a variation of
the `neural analogical reasoning' approach of (Velickovic and Blundell 2021),
we instead search for a sequence of elementary neural network transformations
that manipulate distributed representations derived from a symbolic space, to
which input images are directly encoded. We evaluate the extent to which our
`neural reasoning' approach generalizes for images with unseen shapes and
positions.
Related papers
- Disentangling Visual Priors: Unsupervised Learning of Scene Interpretations with Compositional Autoencoder [0.20718016474717196]
We propose a neurosymbolic architecture that uses a domain-specific language to capture selected priors of image formation.
We express template programs in that language and learn their parameterization with features extracted from the scene by a convolutional neural network.
When executed, the parameterized program produces geometric primitives which are rendered and assessed for correspondence with the scene content.
arXiv Detail & Related papers (2024-09-15T12:47:39Z) - Patch-wise Graph Contrastive Learning for Image Translation [69.85040887753729]
We exploit the graph neural network to capture the topology-aware features.
We construct the graph based on the patch-wise similarity from a pretrained encoder.
In order to capture the hierarchical semantic structure, we propose the graph pooling.
arXiv Detail & Related papers (2023-12-13T15:45:19Z) - On the Transition from Neural Representation to Symbolic Knowledge [2.2528422603742304]
We propose a Neural-Symbolic Transitional Dictionary Learning (TDL) framework that employs an EM algorithm to learn a transitional representation of data.
We implement the framework with a diffusion model by regarding the decomposition of input as a cooperative game.
We additionally use RL enabled by the Markovian of diffusion models to further tune the learned prototypes.
arXiv Detail & Related papers (2023-08-03T19:29:35Z) - Unsupervised Learning of Invariance Transformations [105.54048699217668]
We develop an algorithmic framework for finding approximate graph automorphisms.
We discuss how this framework can be used to find approximate automorphisms in weighted graphs in general.
arXiv Detail & Related papers (2023-07-24T17:03:28Z) - Imaging with Equivariant Deep Learning [9.333799633608345]
We review the emerging field of equivariant imaging and show how it can provide improved generalization and new imaging opportunities.
We show the interplay between the acquisition physics and group actions and links to iterative reconstruction, blind compressed sensing and self-supervised learning.
arXiv Detail & Related papers (2022-09-05T02:13:57Z) - Compositional Sketch Search [91.84489055347585]
We present an algorithm for searching image collections using free-hand sketches.
We exploit drawings as a concise and intuitive representation for specifying entire scene compositions.
arXiv Detail & Related papers (2021-06-15T09:38:09Z) - Self-Supervised Graph Representation Learning via Topology
Transformations [61.870882736758624]
We present the Topology Transformation Equivariant Representation learning, a general paradigm of self-supervised learning for node representations of graph data.
In experiments, we apply the proposed model to the downstream node and graph classification tasks, and results show that the proposed method outperforms the state-of-the-art unsupervised approaches.
arXiv Detail & Related papers (2021-05-25T06:11:03Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - A Flexible Framework for Designing Trainable Priors with Adaptive
Smoothing and Game Encoding [57.1077544780653]
We introduce a general framework for designing and training neural network layers whose forward passes can be interpreted as solving non-smooth convex optimization problems.
We focus on convex games, solved by local agents represented by the nodes of a graph and interacting through regularization functions.
This approach is appealing for solving imaging problems, as it allows the use of classical image priors within deep models that are trainable end to end.
arXiv Detail & Related papers (2020-06-26T08:34:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.