A survey of multimodal deep generative models
- URL: http://arxiv.org/abs/2207.02127v1
- Date: Tue, 5 Jul 2022 15:48:51 GMT
- Title: A survey of multimodal deep generative models
- Authors: Masahiro Suzuki, Yutaka Matsuo
- Abstract summary: Multimodal learning is a framework for building models that make predictions based on different types of modalities.
Deep generative models in which distributions are parameterized by deep neural networks have attracted much attention.
- Score: 20.717591403306287
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multimodal learning is a framework for building models that make predictions
based on different types of modalities. Important challenges in multimodal
learning are the inference of shared representations from arbitrary modalities
and cross-modal generation via these representations; however, achieving this
requires taking the heterogeneous nature of multimodal data into account. In
recent years, deep generative models, i.e., generative models in which
distributions are parameterized by deep neural networks, have attracted much
attention, especially variational autoencoders, which are suitable for
accomplishing the above challenges because they can consider heterogeneity and
infer good representations of data. Therefore, various multimodal generative
models based on variational autoencoders, called multimodal deep generative
models, have been proposed in recent years. In this paper, we provide a
categorized survey of studies on multimodal deep generative models.
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.