AGAR: Attention Graph-RNN for Adaptative Motion Prediction of Point
Clouds of Deformable Objects
- URL: http://arxiv.org/abs/2307.09936v1
- Date: Wed, 19 Jul 2023 12:21:39 GMT
- Title: AGAR: Attention Graph-RNN for Adaptative Motion Prediction of Point
Clouds of Deformable Objects
- Authors: Pedro Gomes, Silvia Rossi, Laura Toni
- Abstract summary: We propose an improved architecture for point cloud prediction of deformable 3D objects.
Specifically, to handle deformable shapes, we propose a graph-based approach that learns and exploits the spatial structure of point clouds.
The proposed adaptative module controls the composition of local and global motions for each point, enabling the network to model complex motions in deformable 3D objects more effectively.
- Score: 7.414594429329531
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper focuses on motion prediction for point cloud sequences in the
challenging case of deformable 3D objects, such as human body motion. First, we
investigate the challenges caused by deformable shapes and complex motions
present in this type of representation, with the ultimate goal of understanding
the technical limitations of state-of-the-art models. From this understanding,
we propose an improved architecture for point cloud prediction of deformable 3D
objects. Specifically, to handle deformable shapes, we propose a graph-based
approach that learns and exploits the spatial structure of point clouds to
extract more representative features. Then we propose a module able to combine
the learned features in an adaptative manner according to the point cloud
movements. The proposed adaptative module controls the composition of local and
global motions for each point, enabling the network to model complex motions in
deformable 3D objects more effectively. We tested the proposed method on the
following datasets: MNIST moving digits, the Mixamo human bodies motions, JPEG
and CWIPC-SXR real-world dynamic bodies. Simulation results demonstrate that
our method outperforms the current baseline methods given its improved ability
to model complex movements as well as preserve point cloud shape. Furthermore,
we demonstrate the generalizability of the proposed framework for dynamic
feature learning, by testing the framework for action recognition on the
MSRAction3D dataset and achieving results on-par with state-of-the-art methods
Related papers
- Dynamic Scene Understanding through Object-Centric Voxelization and Neural Rendering [57.895846642868904]
We present a 3D generative model named DynaVol-S for dynamic scenes that enables object-centric learning.
voxelization infers per-object occupancy probabilities at individual spatial locations.
Our approach integrates 2D semantic features to create 3D semantic grids, representing the scene through multiple disentangled voxel grids.
arXiv Detail & Related papers (2024-07-30T15:33:58Z) - Masked Generative Extractor for Synergistic Representation and 3D Generation of Point Clouds [6.69660410213287]
We propose an innovative framework called Point-MGE to explore the benefits of deeply integrating 3D representation learning and generative learning.
In shape classification, Point-MGE achieved an accuracy of 94.2% (+1.0%) on the ModelNet40 dataset and 92.9% (+5.5%) on the ScanObjectNN dataset.
Experimental results also confirmed that Point-MGE can generate high-quality 3D shapes in both unconditional and conditional settings.
arXiv Detail & Related papers (2024-06-25T07:57:03Z) - Object Dynamics Modeling with Hierarchical Point Cloud-based Representations [1.3934784414106087]
We propose a novel U-net architecture based on continuous point convolution which embeds information from 3D coordinates.
Bottleneck layers in the downsampled point clouds lead to better long-range interaction modeling.
Our approach significantly improves the state-of-the-art, especially in scenarios that require accurate gravity or collision reasoning.
arXiv Detail & Related papers (2024-04-09T06:10:15Z) - Dynamic 3D Point Cloud Sequences as 2D Videos [81.46246338686478]
3D point cloud sequences serve as one of the most common and practical representation modalities of real-world environments.
We propose a novel generic representation called textitStructured Point Cloud Videos (SPCVs)
SPCVs re-organizes a point cloud sequence as a 2D video with spatial smoothness and temporal consistency, where the pixel values correspond to the 3D coordinates of points.
arXiv Detail & Related papers (2024-03-02T08:18:57Z) - Robust 3D Tracking with Quality-Aware Shape Completion [67.9748164949519]
We propose a synthetic target representation composed of dense and complete point clouds depicting the target shape precisely by shape completion for robust 3D tracking.
Specifically, we design a voxelized 3D tracking framework with shape completion, in which we propose a quality-aware shape completion mechanism to alleviate the adverse effect of noisy historical predictions.
arXiv Detail & Related papers (2023-12-17T04:50:24Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - ACID: Action-Conditional Implicit Visual Dynamics for Deformable Object
Manipulation [135.10594078615952]
We introduce ACID, an action-conditional visual dynamics model for volumetric deformable objects.
A benchmark contains over 17,000 action trajectories with six types of plush toys and 78 variants.
Our model achieves the best performance in geometry, correspondence, and dynamics predictions.
arXiv Detail & Related papers (2022-03-14T04:56:55Z) - Self-Supervised Learning of Part Mobility from Point Cloud Sequence [9.495859862104515]
We introduce a self-supervised method for segmenting parts and predicting their motion attributes from a point sequence representing a dynamic object.
We generate trajectories by using correlations among successive frames of the sequence.
We evaluate our method on various tasks including motion part segmentation, motion axis prediction and motion range estimation.
arXiv Detail & Related papers (2020-10-20T11:29:46Z) - InfoFocus: 3D Object Detection for Autonomous Driving with Dynamic
Information Modeling [65.47126868838836]
We propose a novel 3D object detection framework with dynamic information modeling.
Coarse predictions are generated in the first stage via a voxel-based region proposal network.
Experiments are conducted on the large-scale nuScenes 3D detection benchmark.
arXiv Detail & Related papers (2020-07-16T18:27:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.