Explore Contextual Information for 3D Scene Graph Generation
- URL: http://arxiv.org/abs/2210.06240v1
- Date: Wed, 12 Oct 2022 14:26:17 GMT
- Title: Explore Contextual Information for 3D Scene Graph Generation
- Authors: Yuanyuan Liu, Chengjiang Long, Zhaoxuan Zhang, Bokai Liu, Qiang Zhang,
Baocai Yin, Xin Yang
- Abstract summary: 3D scene graph generation (SGG) has been of high interest in computer vision.
We propose a framework fully exploring contextual information for the 3D SGG task.
Our approach achieves superior or competitive performance over previous methods on the 3DSSG dataset.
- Score: 43.66442227874461
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D scene graph generation (SGG) has been of high interest in computer vision.
Although the accuracy of 3D SGG on coarse classification and single relation
label has been gradually improved, the performance of existing works is still
far from being perfect for fine-grained and multi-label situations. In this
paper, we propose a framework fully exploring contextual information for the 3D
SGG task, which attempts to satisfy the requirements of fine-grained entity
class, multiple relation labels, and high accuracy simultaneously. Our proposed
approach is composed of a Graph Feature Extraction module and a Graph
Contextual Reasoning module, achieving appropriate information-redundancy
feature extraction, structured organization, and hierarchical inferring. Our
approach achieves superior or competitive performance over previous methods on
the 3DSSG dataset, especially on the relationship prediction sub-task.
Related papers
- ESGNN: Towards Equivariant Scene Graph Neural Network for 3D Scene Understanding [2.5165775267615205]
This work is the first to implement an Equivariant Graph Neural Network in semantic scene graph generation from 3D point clouds for scene understanding.
Our proposed method, ESGNN, outperforms existing state-of-the-art approaches, demonstrating a significant improvement in scene estimation with faster convergence.
arXiv Detail & Related papers (2024-06-30T06:58:04Z) - Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - GraphRelate3D: Context-Dependent 3D Object Detection with Inter-Object Relationship Graphs [13.071451453118783]
We introduce an object relation module, consisting of a graph generator and a graph neural network (GNN) to learn the spatial information from certain patterns to improve 3D object detection.
Our approach improves upon the baseline PV-RCNN on the KITTI validation set for the car class across easy, moderate, and hard difficulty levels by 0.82%, 0.74%, and 0.58%, respectively.
arXiv Detail & Related papers (2024-05-10T19:18:02Z) - S^2Former-OR: Single-Stage Bi-Modal Transformer for Scene Graph Generation in OR [50.435592120607815]
Scene graph generation (SGG) of surgical procedures is crucial in enhancing holistically cognitive intelligence in the operating room (OR)
Previous works have primarily relied on multi-stage learning, where the generated semantic scene graphs depend on intermediate processes with pose estimation and object detection.
In this study, we introduce a novel single-stage bi-modal transformer framework for SGG in the OR, termed S2Former-OR.
arXiv Detail & Related papers (2024-02-22T11:40:49Z) - SGTR+: End-to-end Scene Graph Generation with Transformer [42.396971149458324]
Scene Graph Generation (SGG) remains a challenging visual understanding task due to its compositional property.
Most previous works adopt a bottom-up, two-stage or point-based, one-stage approach, which often suffers from high time complexity or suboptimal designs.
We propose a novel SGG method to address the aforementioned issues, formulating the task as a bipartite graph construction problem.
arXiv Detail & Related papers (2024-01-23T15:18:20Z) - Incremental 3D Semantic Scene Graph Prediction from RGB Sequences [86.77318031029404]
We propose a real-time framework that incrementally builds a consistent 3D semantic scene graph of a scene given an RGB image sequence.
Our method consists of a novel incremental entity estimation pipeline and a scene graph prediction network.
The proposed network estimates 3D semantic scene graphs with iterative message passing using multi-view and geometric features extracted from the scene entities.
arXiv Detail & Related papers (2023-05-04T11:32:16Z) - Semantic Compositional Learning for Low-shot Scene Graph Generation [122.51930904132685]
Many scene graph generation (SGG) models solely use the limited annotated relation triples for training.
We propose a novel semantic compositional learning strategy that makes it possible to construct additional, realistic relation triples.
For three recent SGG models, adding our strategy improves their performance by close to 50%, and all of them substantially exceed the current state-of-the-art.
arXiv Detail & Related papers (2021-08-19T10:13:55Z) - Learning 3D Semantic Scene Graphs from 3D Indoor Reconstructions [94.17683799712397]
We focus on scene graphs, a data structure that organizes the entities of a scene in a graph.
We propose a learned method that regresses a scene graph from the point cloud of a scene.
We show the application of our method in a domain-agnostic retrieval task, where graphs serve as an intermediate representation for 3D-3D and 2D-3D matching.
arXiv Detail & Related papers (2020-04-08T12:25:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.