Generalized Multi-view Shared Subspace Learning using View Bootstrapping
- URL: http://arxiv.org/abs/2005.06038v1
- Date: Tue, 12 May 2020 20:35:14 GMT
- Title: Generalized Multi-view Shared Subspace Learning using View Bootstrapping
- Authors: Krishna Somandepalli and Shrikanth Narayanan
- Abstract summary: Key objective in multi-view learning is to model the information common to multiple parallel views of a class of objects/events to improve downstream learning tasks.
We present a neural method based on multi-view correlation to capture the information shared across a large number of views by subsampling them in a view-agnostic manner during training.
Experiments on spoken word recognition, 3D object classification and pose-invariant face recognition demonstrate the robustness of view bootstrapping to model a large number of views.
- Score: 43.027427742165095
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A key objective in multi-view learning is to model the information common to
multiple parallel views of a class of objects/events to improve downstream
learning tasks. In this context, two open research questions remain: How can we
model hundreds of views per event? Can we learn robust multi-view embeddings
without any knowledge of how these views are acquired? We present a neural
method based on multi-view correlation to capture the information shared across
a large number of views by subsampling them in a view-agnostic manner during
training. To provide an upper bound on the number of views to subsample for a
given embedding dimension, we analyze the error of the bootstrapped multi-view
correlation objective using matrix concentration theory. Our experiments on
spoken word recognition, 3D object classification and pose-invariant face
recognition demonstrate the robustness of view bootstrapping to model a large
number of views. Results underscore the applicability of our method for a
view-agnostic learning setting.
Related papers
- POV: Prompt-Oriented View-Agnostic Learning for Egocentric Hand-Object
Interaction in the Multi-View World [59.545114016224254]
Humans are good at translating third-person observations of hand-object interactions into an egocentric view.
We propose a Prompt-Oriented View-agnostic learning framework, which enables this view adaptation with few egocentric videos.
arXiv Detail & Related papers (2024-03-09T09:54:44Z) - Multi-view Fuzzy Representation Learning with Rules based Model [25.997490574254172]
Unsupervised multi-view representation learning has been extensively studied for mining multi-view data.
This paper proposes a new multi-view fuzzy representation learning method based on the interpretable Takagi-Sugeno-Kang fuzzy system (MVRL_FS)
arXiv Detail & Related papers (2023-09-20T17:13:15Z) - Learning from Semantic Alignment between Unpaired Multiviews for
Egocentric Video Recognition [23.031934558964473]
We propose Semantics-based Unpaired Multiview Learning (SUM-L) to tackle this unpaired multiview learning problem.
Key idea is to build cross-view pseudo-pairs and do view-invariant alignment by leveraging the semantic information of videos.
Our method also outperforms multiple existing view-alignment methods, under the more challenging scenario.
arXiv Detail & Related papers (2023-08-22T15:10:42Z) - Multi-View Class Incremental Learning [57.14644913531313]
Multi-view learning (MVL) has gained great success in integrating information from multiple perspectives of a dataset to improve downstream task performance.
This paper investigates a novel paradigm called multi-view class incremental learning (MVCIL), where a single model incrementally classifies new classes from a continual stream of views.
arXiv Detail & Related papers (2023-06-16T08:13:41Z) - Investigating and Mitigating the Side Effects of Noisy Views for Self-Supervised Clustering Algorithms in Practical Multi-View Scenarios [35.32285779434823]
Multi-view clustering (MVC) aims at exploring category structures among multi-view data in self-supervised manners.
noisy views might seriously degenerate when the views are noisy in practical multi-view scenarios.
We propose a theoretically grounded deep MVC method (namely MVCAN) to address this issue.
arXiv Detail & Related papers (2023-03-30T09:22:17Z) - Cross-view Graph Contrastive Representation Learning on Partially
Aligned Multi-view Data [52.491074276133325]
Multi-view representation learning has developed rapidly over the past decades and has been applied in many fields.
We propose a new cross-view graph contrastive learning framework, which integrates multi-view information to align data and learn latent representations.
Experiments conducted on several real datasets demonstrate the effectiveness of the proposed method on the clustering and classification tasks.
arXiv Detail & Related papers (2022-11-08T09:19:32Z) - Matching Multiple Perspectives for Efficient Representation Learning [0.0]
We present an approach that combines self-supervised learning with a multi-perspective matching technique.
We show that the availability of multiple views of the same object combined with a variety of self-supervised pretraining algorithms can lead to improved object classification performance.
arXiv Detail & Related papers (2022-08-16T10:33:13Z) - Embedded Deep Bilinear Interactive Information and Selective Fusion for
Multi-view Learning [70.67092105994598]
We propose a novel multi-view learning framework to make the multi-view classification better aimed at the above-mentioned two aspects.
In particular, we train different deep neural networks to learn various intra-view representations.
Experiments on six publicly available datasets demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2020-07-13T01:13:23Z) - Exploit Clues from Views: Self-Supervised and Regularized Learning for
Multiview Object Recognition [66.87417785210772]
This work investigates the problem of multiview self-supervised learning (MV-SSL)
A novel surrogate task for self-supervised learning is proposed by pursuing "object invariant" representation.
Experiments shows that the recognition and retrieval results using view invariant prototype embedding (VISPE) outperform other self-supervised learning methods.
arXiv Detail & Related papers (2020-03-28T07:06:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.