Effective Abstract Reasoning with Dual-Contrast Network
- URL: http://arxiv.org/abs/2205.13720v1
- Date: Fri, 27 May 2022 02:26:52 GMT
- Title: Effective Abstract Reasoning with Dual-Contrast Network
- Authors: Tao Zhuo and Mohan Kankanhalli
- Abstract summary: We aim to solve Raven's Progressive Matrices ( RPM) puzzles with neural networks.
We design a simple yet effective Dual-Contrast Network (DCNet) to exploit the inherent structure of RPM puzzles.
Experimental results on the RAVEN and PGM datasets show that DCNet outperforms the state-of-the-art methods by a large margin of 5.77%.
- Score: 10.675709291797535
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a step towards improving the abstract reasoning capability of machines, we
aim to solve Raven's Progressive Matrices (RPM) with neural networks, since
solving RPM puzzles is highly correlated with human intelligence. Unlike
previous methods that use auxiliary annotations or assume hidden rules to
produce appropriate feature representation, we only use the ground truth answer
of each question for model learning, aiming for an intelligent agent to have a
strong learning capability with a small amount of supervision. Based on the RPM
problem formulation, the correct answer filled into the missing entry of the
third row/column has to best satisfy the same rules shared between the first
two rows/columns. Thus we design a simple yet effective Dual-Contrast Network
(DCNet) to exploit the inherent structure of RPM puzzles. Specifically, a rule
contrast module is designed to compare the latent rules between the filled
row/column and the first two rows/columns; a choice contrast module is designed
to increase the relative differences between candidate choices. Experimental
results on the RAVEN and PGM datasets show that DCNet outperforms the
state-of-the-art methods by a large margin of 5.77%. Further experiments on few
training samples and model generalization also show the effectiveness of DCNet.
Code is available at https://github.com/visiontao/dcnet.
Related papers
- MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Dual Lottery Ticket Hypothesis [71.95937879869334]
Lottery Ticket Hypothesis (LTH) provides a novel view to investigate sparse network training and maintain its capacity.
In this work, we regard the winning ticket from LTH as the subnetwork which is in trainable condition and its performance as our benchmark.
We propose a simple sparse network training strategy, Random Sparse Network Transformation (RST), to substantiate our DLTH.
arXiv Detail & Related papers (2022-03-08T18:06:26Z) - Sketching as a Tool for Understanding and Accelerating Self-attention
for Long Sequences [52.6022911513076]
Transformer-based models are not efficient in processing long sequences due to the quadratic space and time complexity of the self-attention modules.
We propose Linformer and Informer to reduce the quadratic complexity to linear (modulo logarithmic factors) via low-dimensional projection and row selection.
Based on the theoretical analysis, we propose Skeinformer to accelerate self-attention and further improve the accuracy of matrix approximation to self-attention.
arXiv Detail & Related papers (2021-12-10T06:58:05Z) - Unsupervised Abstract Reasoning for Raven's Problem Matrices [9.278113063631643]
Raven's Progressive Matrices ( RPM) is highly correlated with human intelligence.
We propose the first unsupervised learning method for solving RPM problems.
Our method even outperforms some of the supervised approaches.
arXiv Detail & Related papers (2021-09-21T07:44:58Z) - Multi-Model Least Squares-Based Recomputation Framework for Large Data
Analysis [0.0]
In complex tasks such as handling the ImageNet dataset, there are often many more clues that can be directly encoded.
This serves as the motivation to retrain the latent space representations to learn some clues that unsupervised learning has not yet learned.
In this paper, a recomputation-based multilayer network using MP inverse (RML-MP) is developed.
arXiv Detail & Related papers (2021-01-04T23:01:30Z) - Multi-Label Contrastive Learning for Abstract Visual Reasoning [0.0]
State-of-the-art systems solving Raven's Progressive Matrices rely on massive pattern-based training and exploiting biases in the dataset.
Humans concentrate on identification of the rules / concepts underlying the RPM (or generally a visual reasoning task) to be solved.
We propose a new sparse rule encoding scheme for RPMs which, besides the new training algorithm, is the key factor contributing to the state-of-the-art performance.
arXiv Detail & Related papers (2020-12-03T14:18:15Z) - Stratified Rule-Aware Network for Abstract Visual Reasoning [46.015682319351676]
Raven's Progressive Matrices (RPM) test is typically used to examine the capability of abstract reasoning.
Recent studies, taking advantage of Convolutional Neural Networks (CNNs), have achieved encouraging progress to accomplish the RPM test.
We propose a Stratified Rule-Aware Network (SRAN) to generate the rule embeddings for two input sequences.
arXiv Detail & Related papers (2020-02-17T08:44:05Z) - Solving Raven's Progressive Matrices with Neural Networks [10.675709291797535]
Raven's Progressive Matrices (RPM) have been widely used for Intelligence Quotient (IQ) test of humans.
In this paper, we aim to solve RPM with neural networks in both supervised and unsupervised manners.
arXiv Detail & Related papers (2020-02-05T05:18:02Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.