Environment-Invariant Curriculum Relation Learning for Fine-Grained
Scene Graph Generation
- URL: http://arxiv.org/abs/2308.03282v2
- Date: Mon, 21 Aug 2023 01:19:33 GMT
- Title: Environment-Invariant Curriculum Relation Learning for Fine-Grained
Scene Graph Generation
- Authors: Yukuan Min and Aming Wu and Cheng Deng
- Abstract summary: The scene graph generation (SGG) task is designed to identify the predicates based on the subject-object pairs.
We propose a novel Environment Invariant Curriculum Relation learning (EICR) method, which can be applied in a plug-and-play fashion to existing SGG methods.
- Score: 66.62453697902947
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The scene graph generation (SGG) task is designed to identify the predicates
based on the subject-object pairs.However,existing datasets generally include
two imbalance cases: one is the class imbalance from the predicted predicates
and another is the context imbalance from the given subject-object pairs, which
presents significant challenges for SGG. Most existing methods focus on the
imbalance of the predicted predicate while ignoring the imbalance of the
subject-object pairs, which could not achieve satisfactory results. To address
the two imbalance cases, we propose a novel Environment Invariant Curriculum
Relation learning (EICR) method, which can be applied in a plug-and-play
fashion to existing SGG methods. Concretely, to remove the imbalance of the
subject-object pairs, we first construct different distribution environments
for the subject-object pairs and learn a model invariant to the environment
changes. Then, we construct a class-balanced curriculum learning strategy to
balance the different environments to remove the predicate imbalance.
Comprehensive experiments conducted on VG and GQA datasets demonstrate that our
EICR framework can be taken as a general strategy for various SGG models, and
achieve significant improvements.
Related papers
- Semantic Diversity-aware Prototype-based Learning for Unbiased Scene Graph Generation [21.772806350802203]
In scene graph generation (SGG) datasets, each subject-object pair is annotated with a single predicate.
Existing SGG models are trained to predict the one and only predicate for each pair.
This in turn results in the SGG models to overlook the semantic diversity that may exist in a predicate.
arXiv Detail & Related papers (2024-07-22T05:53:46Z) - TD^2-Net: Toward Denoising and Debiasing for Dynamic Scene Graph
Generation [76.24766055944554]
We introduce a network named TD$2$-Net that aims at denoising and debiasing for dynamic SGG.
TD$2$-Net outperforms the second-best competitors by 12.7 % on mean-Recall@10 for predicate classification.
arXiv Detail & Related papers (2024-01-23T04:17:42Z) - Adaptive Weighted Co-Learning for Cross-Domain Few-Shot Learning [23.615250207134004]
Cross-domain few-shot learning (CDFSL) induces a very challenging adaptation problem.
We propose a simple Adaptive Weighted Co-Learning (AWCoL) method to address the CDFSL challenge.
Comprehensive experiments are conducted on multiple benchmark datasets and the empirical results demonstrate that the proposed method produces state-of-the-art CDFSL performance.
arXiv Detail & Related papers (2023-12-06T22:09:52Z) - Rethinking Semi-Supervised Imbalanced Node Classification from
Bias-Variance Decomposition [18.3055496602884]
This paper introduces a new approach to address the issue of class imbalance in graph neural networks (GNNs) for learning on graph-structured data.
Our approach integrates imbalanced node classification and Bias-Variance Decomposition, establishing a theoretical framework that closely relates data imbalance to model variance.
arXiv Detail & Related papers (2023-10-28T17:28:07Z) - Generalized Unbiased Scene Graph Generation [85.22334551067617]
Generalized Unbiased Scene Graph Generation (G-USGG) takes into account both predicate-level and concept-level imbalance.
We propose the Multi-Concept Learning (MCL) framework, which ensures a balanced learning process across rare/ uncommon/ common concepts.
arXiv Detail & Related papers (2023-08-09T08:51:03Z) - Delving into Identify-Emphasize Paradigm for Combating Unknown Bias [52.76758938921129]
We propose an effective bias-conflicting scoring method (ECS) to boost the identification accuracy.
We also propose gradient alignment (GA) to balance the contributions of the mined bias-aligned and bias-conflicting samples.
Experiments are conducted on multiple datasets in various settings, demonstrating that the proposed solution can mitigate the impact of unknown biases.
arXiv Detail & Related papers (2023-02-22T14:50:24Z) - Towards Open-vocabulary Scene Graph Generation with Prompt-based
Finetuning [84.39787427288525]
Scene graph generation (SGG) is a fundamental task aimed at detecting visual relations between objects in an image.
We introduce open-vocabulary scene graph generation, a novel, realistic and challenging setting in which a model is trained on a set of base object classes.
Our method can support inference over completely unseen object classes, which existing methods are incapable of handling.
arXiv Detail & Related papers (2022-08-17T09:05:38Z) - Adaptive Fine-Grained Predicates Learning for Scene Graph Generation [122.4588401267544]
General Scene Graph Generation (SGG) models tend to predict head predicates and re-balancing strategies prefer tail categories.
We propose an Adaptive Fine-Grained Predicates Learning (FGPL-A) which aims at differentiating hard-to-distinguish predicates for SGG.
Our proposed model-agnostic strategy significantly boosts performance of benchmark models on VG-SGG and GQA-SGG datasets by up to 175% and 76% on Mean Recall@100, achieving new state-of-the-art performance.
arXiv Detail & Related papers (2022-07-11T03:37:57Z) - Supervised Contrastive Learning for Pre-trained Language Model
Fine-tuning [23.00300794016583]
State-of-the-art natural language understanding classification models follow two-stages.
We propose a supervised contrastive learning (SCL) objective for the fine-tuning stage.
Our proposed fine-tuning objective leads to models that are more robust to different levels of noise in the fine-tuning training data.
arXiv Detail & Related papers (2020-11-03T01:10:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.