Multi-view mid fusion: a universal approach for learning in an HDLSS setting
- URL: http://arxiv.org/abs/2507.06026v1
- Date: Tue, 08 Jul 2025 14:31:53 GMT
- Title: Multi-view mid fusion: a universal approach for learning in an HDLSS setting
- Authors: Lynn Houthuys,
- Abstract summary: This paper introduces a universal approach for learning in HDLSS setting using multi-view mid fusion techniques.<n>It shows how existing mid fusion multi-view methods perform well in an HDLSS setting even if no inherent views are provided.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The high-dimensional low-sample-size (HDLSS) setting presents significant challenges in various applications where the feature dimension far exceeds the number of available samples. This paper introduces a universal approach for learning in HDLSS setting using multi-view mid fusion techniques. It shows how existing mid fusion multi-view methods perform well in an HDLSS setting even if no inherent views are provided. Three view construction methods are proposed that split the high-dimensional feature vectors into smaller subsets, each representing a different view. Extensive experimental validation across model-types and learning tasks confirm the effectiveness and generalization of the approach. We believe the work in this paper lays the foundation for further research into the universal benefits of multi-view mid fusion learning.
Related papers
- Wasserstein-Aligned Hyperbolic Multi-View Clustering [58.29261653100388]
This paper proposes a novel Wasserstein-Aligned Hyperbolic (WAH) framework for multi-view clustering.<n>Our method exploits a view-specific hyperbolic encoder for each view to embed features into the Lorentz manifold for hierarchical semantic modeling.
arXiv Detail & Related papers (2025-12-10T07:56:19Z) - Adaptive Weighted LSSVM for Multi-View Classification [0.5161531917413708]
AW-LSSVM promotes complementary learning by an iterative global coupling to make each view focus on hard samples of others from previous iterations.<n>Experiments demonstrate AW-LSSVM outperforms existing kernel-based multi-view methods on most datasets.
arXiv Detail & Related papers (2025-12-02T11:14:47Z) - Multiview Manifold Evidential Fusion for PolSAR Image Classification [51.41332458376411]
We propose a new framework to integrate PolSAR manifold learning and evidence fusion into a unified architecture.<n>Experiments on three real-world PolSAR datasets demonstrate that the proposed method consistently outperforms existing approaches in accuracy, robustness, and interpretability.
arXiv Detail & Related papers (2025-10-13T09:05:51Z) - Balanced Multi-view Clustering [56.17836963920012]
Multi-view clustering (MvC) aims to integrate information from different views to enhance the capability of the model in capturing the underlying data structures.<n>The widely used joint training paradigm in MvC is potentially not fully leverage the multi-view information.<n>We propose a novel balanced multi-view clustering (BMvC) method, which introduces a view-specific contrastive regularization (VCR) to modulate the optimization of each view.
arXiv Detail & Related papers (2025-01-05T14:42:47Z) - Beyond Mask: Rethinking Guidance Types in Few-shot Segmentation [67.35274834837064]
We develop a universal vision-language framework (UniFSS) to integrate prompts from text, mask, box, and image.
UniFSS significantly outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2024-07-16T08:41:01Z) - Multi-view Fuzzy Representation Learning with Rules based Model [25.997490574254172]
Unsupervised multi-view representation learning has been extensively studied for mining multi-view data.
This paper proposes a new multi-view fuzzy representation learning method based on the interpretable Takagi-Sugeno-Kang fuzzy system (MVRL_FS)
arXiv Detail & Related papers (2023-09-20T17:13:15Z) - TCGF: A unified tensorized consensus graph framework for multi-view
representation learning [27.23929515170454]
This paper proposes a universal multi-view representation learning framework named Consensus Graph Framework (TCGF)
It first provides a unified framework for existing multi-view works to exploit the representations for individual view.
Then, stacks them into a tensor under alignment basics as a high-order representation, allowing for the smooth propagation of consistency.
arXiv Detail & Related papers (2023-09-14T19:29:14Z) - MLF-DET: Multi-Level Fusion for Cross-Modal 3D Object Detection [54.52102265418295]
We propose a novel and effective Multi-Level Fusion network, named as MLF-DET, for high-performance cross-modal 3D object DETection.
For the feature-level fusion, we present the Multi-scale Voxel Image fusion (MVI) module, which densely aligns multi-scale voxel features with image features.
For the decision-level fusion, we propose the lightweight Feature-cued Confidence Rectification (FCR) module, which exploits image semantics to rectify the confidence of detection candidates.
arXiv Detail & Related papers (2023-07-18T11:26:02Z) - Multi-View Class Incremental Learning [57.14644913531313]
Multi-view learning (MVL) has gained great success in integrating information from multiple perspectives of a dataset to improve downstream task performance.
This paper investigates a novel paradigm called multi-view class incremental learning (MVCIL), where a single model incrementally classifies new classes from a continual stream of views.
arXiv Detail & Related papers (2023-06-16T08:13:41Z) - Cross-view Graph Contrastive Representation Learning on Partially
Aligned Multi-view Data [52.491074276133325]
Multi-view representation learning has developed rapidly over the past decades and has been applied in many fields.
We propose a new cross-view graph contrastive learning framework, which integrates multi-view information to align data and learn latent representations.
Experiments conducted on several real datasets demonstrate the effectiveness of the proposed method on the clustering and classification tasks.
arXiv Detail & Related papers (2022-11-08T09:19:32Z) - Multi-view Subspace Adaptive Learning via Autoencoder and Attention [3.8574404853067215]
We propose a new Multiview Subspace Adaptive Learning based on Attention and Autoencoder (MSALAA)
This method combines a deep autoencoder and a method for aligning the self-representations of various views.
We empirically observe significant improvement over existing baseline methods on six real-life datasets.
arXiv Detail & Related papers (2022-01-01T11:31:52Z) - Embedded Deep Bilinear Interactive Information and Selective Fusion for
Multi-view Learning [70.67092105994598]
We propose a novel multi-view learning framework to make the multi-view classification better aimed at the above-mentioned two aspects.
In particular, we train different deep neural networks to learn various intra-view representations.
Experiments on six publicly available datasets demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2020-07-13T01:13:23Z) - Multi-view Low-rank Preserving Embedding: A Novel Method for Multi-view
Representation [11.91574721055601]
This paper proposes a novel multi-view learning method, named Multi-view Low-rank Preserving Embedding (MvLPE)
It integrates different views into one centroid view by minimizing the disagreement term, based on distance or similarity matrix among instances.
Experiments on six benchmark datasets demonstrate that the proposed method outperforms its counterparts.
arXiv Detail & Related papers (2020-06-14T12:47:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.