Provably Learning Object-Centric Representations
- URL: http://arxiv.org/abs/2305.14229v1
- Date: Tue, 23 May 2023 16:44:49 GMT
- Title: Provably Learning Object-Centric Representations
- Authors: Jack Brady, Roland S. Zimmermann, Yash Sharma, Bernhard Sch\"olkopf,
Julius von K\"ugelgen, Wieland Brendel
- Abstract summary: We analyze when object-centric representations can provably be learned without supervision.
We prove that the ground-truth object representations can be identified by an invertible and compositional inference model.
We provide evidence that our theory holds predictive power for existing object-centric models.
- Score: 25.152680199034215
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning structured representations of the visual world in terms of objects
promises to significantly improve the generalization abilities of current
machine learning models. While recent efforts to this end have shown promising
empirical progress, a theoretical account of when unsupervised object-centric
representation learning is possible is still lacking. Consequently,
understanding the reasons for the success of existing object-centric methods as
well as designing new theoretically grounded methods remains challenging. In
the present work, we analyze when object-centric representations can provably
be learned without supervision. To this end, we first introduce two assumptions
on the generative process for scenes comprised of several objects, which we
call compositionality and irreducibility. Under this generative process, we
prove that the ground-truth object representations can be identified by an
invertible and compositional inference model, even in the presence of
dependencies between objects. We empirically validate our results through
experiments on synthetic data. Finally, we provide evidence that our theory
holds predictive power for existing object-centric models by showing a close
correspondence between models' compositionality and invertibility and their
empirical identifiability.
Related papers
- Zero-Shot Object-Centric Representation Learning [72.43369950684057]
We study current object-centric methods through the lens of zero-shot generalization.
We introduce a benchmark comprising eight different synthetic and real-world datasets.
We find that training on diverse real-world images improves transferability to unseen scenarios.
arXiv Detail & Related papers (2024-08-17T10:37:07Z) - Learning Discrete Concepts in Latent Hierarchical Models [73.01229236386148]
Learning concepts from natural high-dimensional data holds potential in building human-aligned and interpretable machine learning models.
We formalize concepts as discrete latent causal variables that are related via a hierarchical causal model.
We substantiate our theoretical claims with synthetic data experiments.
arXiv Detail & Related papers (2024-06-01T18:01:03Z) - Learning Interpretable Concepts: Unifying Causal Representation Learning
and Foundation Models [51.43538150982291]
We study how to learn human-interpretable concepts from data.
Weaving together ideas from both fields, we show that concepts can be provably recovered from diverse data.
arXiv Detail & Related papers (2024-02-14T15:23:59Z) - Robust and Controllable Object-Centric Learning through Energy-based
Models [95.68748828339059]
ours is a conceptually simple and general approach to learning object-centric representations through an energy-based model.
We show that ours can be easily integrated into existing architectures and can effectively extract high-quality object-centric representations.
arXiv Detail & Related papers (2022-10-11T15:11:15Z) - Bridging the Gap to Real-World Object-Centric Learning [66.55867830853803]
We show that reconstructing features from models trained in a self-supervised manner is a sufficient training signal for object-centric representations to arise in a fully unsupervised way.
Our approach, DINOSAUR, significantly out-performs existing object-centric learning models on simulated data.
arXiv Detail & Related papers (2022-09-29T15:24:47Z) - Causal Reasoning Meets Visual Representation Learning: A Prospective
Study [117.08431221482638]
Lack of interpretability, robustness, and out-of-distribution generalization are becoming the challenges of the existing visual models.
Inspired by the strong inference ability of human-level agents, recent years have witnessed great effort in developing causal reasoning paradigms.
This paper aims to provide a comprehensive overview of this emerging field, attract attention, encourage discussions, bring to the forefront the urgency of developing novel causal reasoning methods.
arXiv Detail & Related papers (2022-04-26T02:22:28Z) - Object Pursuit: Building a Space of Objects via Discriminative Weight
Generation [23.85039747700698]
We propose a framework to continuously learn object-centric representations for visual learning and understanding.
We leverage interactions to sample diverse variations of an object and the corresponding training signals while learning the object-centric representations.
We perform an extensive study of the key features of the proposed framework and analyze the characteristics of the learned representations.
arXiv Detail & Related papers (2021-12-15T08:25:30Z) - Structured World Belief for Reinforcement Learning in POMDP [16.5072728047837]
We propose Structured World Belief, a model for learning and inference of object-centric belief states.
In experiments, we show that object-centric belief provides a more accurate and robust performance for filtering and generation.
arXiv Detail & Related papers (2021-07-19T01:47:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.