Debunking Free Fusion Myth: Online Multi-view Anomaly Detection with
Disentangled Product-of-Experts Modeling
- URL: http://arxiv.org/abs/2310.18728v2
- Date: Tue, 31 Oct 2023 22:52:08 GMT
- Title: Debunking Free Fusion Myth: Online Multi-view Anomaly Detection with
Disentangled Product-of-Experts Modeling
- Authors: Hao Wang, Zhi-Qi Cheng, Jingdong Sun, Xin Yang, Xiao Wu, Hongyang
Chen, and Yan Yang
- Abstract summary: Multi-view or even multi-modal data is appealing yet challenging for real-world applications.
We propose dPoE, a novel multi-view variational autoencoder model that involves (1) a Product-of-Experts layer in tackling multi-view data, (2) a Total Correction discriminator in disentangling view-common and view-specific representations, and (3) a joint loss function in wrapping up all components.
- Score: 25.02446577349165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-view or even multi-modal data is appealing yet challenging for
real-world applications. Detecting anomalies in multi-view data is a prominent
recent research topic. However, most of the existing methods 1) are only
suitable for two views or type-specific anomalies, 2) suffer from the issue of
fusion disentanglement, and 3) do not support online detection after model
deployment. To address these challenges, our main ideas in this paper are
three-fold: multi-view learning, disentangled representation learning, and
generative model. To this end, we propose dPoE, a novel multi-view variational
autoencoder model that involves (1) a Product-of-Experts (PoE) layer in
tackling multi-view data, (2) a Total Correction (TC) discriminator in
disentangling view-common and view-specific representations, and (3) a joint
loss function in wrapping up all components. In addition, we devise theoretical
information bounds to control both view-common and view-specific
representations. Extensive experiments on six real-world datasets markedly
demonstrate that the proposed dPoE outperforms baselines.
Related papers
- SpecRaGE: Robust and Generalizable Multi-view Spectral Representation Learning [9.393841121141076]
Multi-view representation learning (MvRL) has garnered substantial attention in recent years.
graph Laplacian-based MvRL methods have demonstrated remarkable success in representing multi-view data.
We introduce $textitSpecRaGE$, a novel fusion-based framework that integrates the strengths of graph Laplacian methods with the power of deep learning.
arXiv Detail & Related papers (2024-11-04T14:51:35Z) - PoIFusion: Multi-Modal 3D Object Detection via Fusion at Points of Interest [65.48057241587398]
PoIFusion is a framework to fuse information of RGB images and LiDAR point clouds at the points of interest (PoIs)
Our approach maintains the view of each modality and obtains multi-modal features by computation-friendly projection and computation.
We conducted extensive experiments on nuScenes and Argoverse2 datasets to evaluate our approach.
arXiv Detail & Related papers (2024-03-14T09:28:12Z) - Multi-view Fuzzy Representation Learning with Rules based Model [25.997490574254172]
Unsupervised multi-view representation learning has been extensively studied for mining multi-view data.
This paper proposes a new multi-view fuzzy representation learning method based on the interpretable Takagi-Sugeno-Kang fuzzy system (MVRL_FS)
arXiv Detail & Related papers (2023-09-20T17:13:15Z) - Deep Incomplete Multi-view Clustering with Cross-view Partial Sample and
Prototype Alignment [50.82982601256481]
We propose a Cross-view Partial Sample and Prototype Alignment Network (CPSPAN) for Deep Incomplete Multi-view Clustering.
Unlike existing contrastive-based methods, we adopt pair-observed data alignment as 'proxy supervised signals' to guide instance-to-instance correspondence construction.
arXiv Detail & Related papers (2023-03-28T02:31:57Z) - Credible Remote Sensing Scene Classification Using Evidential Fusion on
Aerial-Ground Dual-view Images [6.817740582240199]
Multi-view (multi-source, multi-modal, multi-perspective, etc.) data are being used more frequently in remote sensing tasks.
The issue of data quality becomes more apparent, limiting the potential benefits of multi-view data.
Deep learning is introduced to the task of aerial-ground dual-view remote sensing scene classification.
arXiv Detail & Related papers (2023-01-02T12:27:55Z) - Progressive Multi-view Human Mesh Recovery with Self-Supervision [68.60019434498703]
Existing solutions typically suffer from poor generalization performance to new settings.
We propose a novel simulation-based training pipeline for multi-view human mesh recovery.
arXiv Detail & Related papers (2022-12-10T06:28:29Z) - Cross-view Graph Contrastive Representation Learning on Partially
Aligned Multi-view Data [52.491074276133325]
Multi-view representation learning has developed rapidly over the past decades and has been applied in many fields.
We propose a new cross-view graph contrastive learning framework, which integrates multi-view information to align data and learn latent representations.
Experiments conducted on several real datasets demonstrate the effectiveness of the proposed method on the clustering and classification tasks.
arXiv Detail & Related papers (2022-11-08T09:19:32Z) - Two-level Data Augmentation for Calibrated Multi-view Detection [51.5746691103591]
We introduce a new multi-view data augmentation pipeline that preserves alignment among views.
We also propose a second level of augmentation applied directly at the scene level.
When combined with our simple multi-view detection model, our two-level augmentation pipeline outperforms all existing baselines.
arXiv Detail & Related papers (2022-10-19T17:55:13Z) - TSK Fuzzy System Towards Few Labeled Incomplete Multi-View Data
Classification [24.01191516774655]
A transductive semi-supervised incomplete multi-view TSK fuzzy system modeling method (SSIMV_TSK) is proposed to address these challenges.
The proposed method integrates missing view imputation, pseudo label learning of unlabeled data, and fuzzy system modeling into a single process to yield a model with interpretable fuzzy rules.
Experimental results on real datasets show that the proposed method significantly outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2021-10-08T11:41:06Z) - Generative Partial Multi-View Clustering [133.36721417531734]
We propose a generative partial multi-view clustering model, named as GP-MVC, to address the incomplete multi-view problem.
First, multi-view encoder networks are trained to learn common low-dimensional representations, followed by a clustering layer to capture the consistent cluster structure across multiple views.
Second, view-specific generative adversarial networks are developed to generate the missing data of one view conditioning on the shared representation given by other views.
arXiv Detail & Related papers (2020-03-29T17:48:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.