Disentangling Multi-view Representations Beyond Inductive Bias
- URL: http://arxiv.org/abs/2308.01634v2
- Date: Fri, 4 Aug 2023 13:22:08 GMT
- Title: Disentangling Multi-view Representations Beyond Inductive Bias
- Authors: Guanzhou Ke, Yang Yu, Guoqing Chao, Xiaoli Wang, Chenyang Xu, and
Shengfeng He
- Abstract summary: We propose a novel multi-view representation disentangling method that ensures both interpretability and generalizability of the resulting representations.
Our experiments on four multi-view datasets demonstrate that our proposed method outperforms 12 comparison methods in terms of clustering and classification performance.
- Score: 32.15900989696017
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-view (or -modality) representation learning aims to understand the
relationships between different view representations. Existing methods
disentangle multi-view representations into consistent and view-specific
representations by introducing strong inductive biases, which can limit their
generalization ability. In this paper, we propose a novel multi-view
representation disentangling method that aims to go beyond inductive biases,
ensuring both interpretability and generalizability of the resulting
representations. Our method is based on the observation that discovering
multi-view consistency in advance can determine the disentangling information
boundary, leading to a decoupled learning objective. We also found that the
consistency can be easily extracted by maximizing the transformation invariance
and clustering consistency between views. These observations drive us to
propose a two-stage framework. In the first stage, we obtain multi-view
consistency by training a consistent encoder to produce semantically-consistent
representations across views as well as their corresponding pseudo-labels. In
the second stage, we disentangle specificity from comprehensive representations
by minimizing the upper bound of mutual information between consistent and
comprehensive representations. Finally, we reconstruct the original data by
concatenating pseudo-labels and view-specific representations. Our experiments
on four multi-view datasets demonstrate that our proposed method outperforms 12
comparison methods in terms of clustering and classification performance. The
visualization results also show that the extracted consistency and specificity
are compact and interpretable. Our code can be found at
\url{https://github.com/Guanzhou-Ke/DMRIB}.
Related papers
- Rethinking Multi-view Representation Learning via Distilled Disentangling [34.14711778177439]
Multi-view representation learning aims to derive robust representations that are both view-consistent and view-specific from diverse data sources.
This paper presents an in-depth analysis of existing approaches in this domain, highlighting the redundancy between view-consistent and view-specific representations.
We propose an innovative framework for multi-view representation learning, which incorporates a technique we term 'distilled disentangling'
arXiv Detail & Related papers (2024-03-16T11:21:24Z) - DealMVC: Dual Contrastive Calibration for Multi-view Clustering [78.54355167448614]
We propose a novel Dual contrastive calibration network for Multi-View Clustering (DealMVC)
We first design a fusion mechanism to obtain a global cross-view feature. Then, a global contrastive calibration loss is proposed by aligning the view feature similarity graph and the high-confidence pseudo-label graph.
During the training procedure, the interacted cross-view feature is jointly optimized at both local and global levels.
arXiv Detail & Related papers (2023-08-17T14:14:28Z) - A Clustering-guided Contrastive Fusion for Multi-view Representation
Learning [7.630965478083513]
We propose a deep fusion network to fuse view-specific representations into the view-common representation.
We also design an asymmetrical contrastive strategy that aligns the view-common representation and each view-specific representation.
In the incomplete view scenario, our proposed method resists noise interference better than those of our competitors.
arXiv Detail & Related papers (2022-12-28T07:21:05Z) - Cross-view Graph Contrastive Representation Learning on Partially
Aligned Multi-view Data [52.491074276133325]
Multi-view representation learning has developed rapidly over the past decades and has been applied in many fields.
We propose a new cross-view graph contrastive learning framework, which integrates multi-view information to align data and learn latent representations.
Experiments conducted on several real datasets demonstrate the effectiveness of the proposed method on the clustering and classification tasks.
arXiv Detail & Related papers (2022-11-08T09:19:32Z) - Dual Representation Learning for One-Step Clustering of Multi-View Data [30.131568561100817]
We propose a novel one-step multi-view clustering method by exploiting the dual representation of both the common and specific information of different views.
With this framework, the representation learning and clustering partition mutually benefit each other, which effectively improve the clustering performance.
arXiv Detail & Related papers (2022-08-30T14:20:26Z) - MORI-RAN: Multi-view Robust Representation Learning via Hybrid
Contrastive Fusion [4.36488705757229]
Multi-view representation learning is essential for many multi-view tasks, such as clustering and classification.
We propose a hybrid contrastive fusion algorithm to extract robust view-common representation from unlabeled data.
Experimental results demonstrated that the proposed method outperforms 12 competitive multi-view methods on four real-world datasets.
arXiv Detail & Related papers (2022-08-26T09:58:37Z) - Variational Distillation for Multi-View Learning [104.17551354374821]
We design several variational information bottlenecks to exploit two key characteristics for multi-view representation learning.
Under rigorously theoretical guarantee, our approach enables IB to grasp the intrinsic correlation between observations and semantic labels.
arXiv Detail & Related papers (2022-06-20T03:09:46Z) - Cross-Modal Discrete Representation Learning [73.68393416984618]
We present a self-supervised learning framework that learns a representation that captures finer levels of granularity across different modalities.
Our framework relies on a discretized embedding space created via vector quantization that is shared across different modalities.
arXiv Detail & Related papers (2021-06-10T00:23:33Z) - Trusted Multi-View Classification [76.73585034192894]
We propose a novel multi-view classification method, termed trusted multi-view classification.
It provides a new paradigm for multi-view learning by dynamically integrating different views at an evidence level.
The proposed algorithm jointly utilizes multiple views to promote both classification reliability and robustness.
arXiv Detail & Related papers (2021-02-03T13:30:26Z) - Deep Partial Multi-View Learning [94.39367390062831]
We propose a novel framework termed Cross Partial Multi-View Networks (CPM-Nets)
We fifirst provide a formal defifinition of completeness and versatility for multi-view representation.
We then theoretically prove the versatility of the learned latent representations.
arXiv Detail & Related papers (2020-11-12T02:29:29Z) - Random Forest for Dissimilarity-based Multi-view Learning [8.185807285320553]
We show that the Random Forest proximity measure can be used to build the dissimilarity representations.
We then propose a Dynamic View Selection method to better combine the view-specific dissimilarity representations.
arXiv Detail & Related papers (2020-07-16T14:52:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.