PartAfford: Part-level Affordance Discovery from 3D Objects
- URL: http://arxiv.org/abs/2202.13519v1
- Date: Mon, 28 Feb 2022 02:58:36 GMT
- Title: PartAfford: Part-level Affordance Discovery from 3D Objects
- Authors: Chao Xu, Yixin Chen, He Wang, Song-Chun Zhu, Yixin Zhu, Siyuan Huang
- Abstract summary: We present a new task of part-level affordance discovery (PartAfford)
Given only the affordance labels per object, the machine is tasked to (i) decompose 3D shapes into parts and (ii) discover how each part corresponds to a certain affordance category.
We propose a novel learning framework for PartAfford, which discovers part-level representations by leveraging only the affordance set supervision and geometric primitive regularization.
- Score: 113.91774531972855
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding what objects could furnish for humans-namely, learning object
affordance-is the crux to bridge perception and action. In the vision
community, prior work primarily focuses on learning object affordance with
dense (e.g., at a per-pixel level) supervision. In stark contrast, we humans
learn the object affordance without dense labels. As such, the fundamental
question to devise a computational model is: What is the natural way to learn
the object affordance from visual appearance and geometry with humanlike sparse
supervision? In this work, we present a new task of part-level affordance
discovery (PartAfford): Given only the affordance labels per object, the
machine is tasked to (i) decompose 3D shapes into parts and (ii) discover how
each part of the object corresponds to a certain affordance category. We
propose a novel learning framework for PartAfford, which discovers part-level
representations by leveraging only the affordance set supervision and geometric
primitive regularization, without dense supervision. The proposed approach
consists of two main components: (i) an abstraction encoder with slot attention
for unsupervised clustering and abstraction, and (ii) an affordance decoder
with branches for part reconstruction, affordance prediction, and cuboidal
primitive regularization. To learn and evaluate PartAfford, we construct a
part-level, cross-category 3D object affordance dataset, annotated with 24
affordance categories shared among >25, 000 objects. We demonstrate that our
method enables both the abstraction of 3D objects and part-level affordance
discovery, with generalizability to difficult and cross-category examples.
Further ablations reveal the contribution of each component.
Related papers
- Going Denser with Open-Vocabulary Part Segmentation [38.395986723880505]
We propose a detector with the ability to predict both open-vocabulary objects and their part segmentation.
This ability comes from two designs. First, we train the detector on the joint of part-level, object-level and image-level data to build the multi-granularity alignment between language and image.
Second, we parse the novel object into its parts by its dense semantic correspondence with the base object.
arXiv Detail & Related papers (2023-05-18T17:59:10Z) - Self-Supervised Visual Representation Learning with Semantic Grouping [50.14703605659837]
We tackle the problem of learning visual representations from unlabeled scene-centric data.
We propose contrastive learning from data-driven semantic slots, namely SlotCon, for joint semantic grouping and representation learning.
arXiv Detail & Related papers (2022-05-30T17:50:59Z) - Self-Supervised Learning of Object Parts for Semantic Segmentation [7.99536002595393]
We argue that self-supervised learning of object parts is a solution to this issue.
Our method surpasses the state-of-the-art on three semantic segmentation benchmarks by 17%-3%.
arXiv Detail & Related papers (2022-04-27T17:55:17Z) - 3D Compositional Zero-shot Learning with DeCompositional Consensus [102.7571947144639]
We argue that part knowledge should be composable beyond the observed object classes.
We present 3D Compositional Zero-shot Learning as a problem of part generalization from seen to unseen object classes.
arXiv Detail & Related papers (2021-11-29T16:34:53Z) - Unsupervised Part Discovery from Contrastive Reconstruction [90.88501867321573]
The goal of self-supervised visual representation learning is to learn strong, transferable image representations.
We propose an unsupervised approach to object part discovery and segmentation.
Our method yields semantic parts consistent across fine-grained but visually distinct categories.
arXiv Detail & Related papers (2021-11-11T17:59:42Z) - 3D AffordanceNet: A Benchmark for Visual Object Affordance Understanding [33.68455617113953]
We present a 3D AffordanceNet dataset, a benchmark of 23k shapes from 23 semantic object categories, annotated with 18 visual affordance categories.
Three state-of-the-art point cloud deep learning networks are evaluated on all tasks.
arXiv Detail & Related papers (2021-03-30T14:46:27Z) - Fine-Grained 3D Shape Classification with Hierarchical Part-View
Attentions [70.0171362989609]
We propose a novel fine-grained 3D shape classification method named FG3D-Net to capture the fine-grained local details of 3D shapes from multiple rendered views.
Our results under the fine-grained 3D shape dataset show that our method outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2020-05-26T06:53:19Z) - A Deep Learning Approach to Object Affordance Segmentation [31.221897360610114]
We design an autoencoder that infers pixel-wise affordance labels in both videos and static images.
Our model surpasses the need for object labels and bounding boxes by using a soft-attention mechanism.
We show that our model achieves competitive results compared to strongly supervised methods on SOR3D-AFF.
arXiv Detail & Related papers (2020-04-18T15:34:41Z) - Global-Local Bidirectional Reasoning for Unsupervised Representation
Learning of 3D Point Clouds [109.0016923028653]
We learn point cloud representation by bidirectional reasoning between the local structures and the global shape without human supervision.
We show that our unsupervised model surpasses the state-of-the-art supervised methods on both synthetic and real-world 3D object classification datasets.
arXiv Detail & Related papers (2020-03-29T08:26:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.