Uncertainty-Aware Multi-View Representation Learning
- URL: http://arxiv.org/abs/2201.05776v1
- Date: Sat, 15 Jan 2022 07:16:20 GMT
- Title: Uncertainty-Aware Multi-View Representation Learning
- Authors: Yu Geng, Zongbo Han, Changqing Zhang, Qinghua Hu
- Abstract summary: We devise a novel unsupervised multi-view learning approach, termed as Dynamic Uncertainty-Aware Networks (DUA-Nets)
Guided by the uncertainty of data estimated from the generation perspective, intrinsic information from multiple views is integrated to obtain noise-free representations.
Our model achieves superior performance in extensive experiments and shows the robustness to noisy data.
- Score: 53.06828186507994
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning from different data views by exploring the underlying complementary
information among them can endow the representation with stronger expressive
ability. However, high-dimensional features tend to contain noise, and
furthermore, the quality of data usually varies for different samples (even for
different views), i.e., one view may be informative for one sample but not the
case for another. Therefore, it is quite challenging to integrate multi-view
noisy data under unsupervised setting. Traditional multi-view methods either
simply treat each view with equal importance or tune the weights of different
views to fixed values, which are insufficient to capture the dynamic noise in
multi-view data. In this work, we devise a novel unsupervised multi-view
learning approach, termed as Dynamic Uncertainty-Aware Networks (DUA-Nets).
Guided by the uncertainty of data estimated from the generation perspective,
intrinsic information from multiple views is integrated to obtain noise-free
representations. Under the help of uncertainty, DUA-Nets weigh each view of
individual sample according to data quality so that the high-quality samples
(or views) can be fully exploited while the effects from the noisy samples (or
views) will be alleviated. Our model achieves superior performance in extensive
experiments and shows the robustness to noisy data.
Related papers
- Regularized Contrastive Partial Multi-view Outlier Detection [76.77036536484114]
We propose a novel method named Regularized Contrastive Partial Multi-view Outlier Detection (RCPMOD)
In this framework, we utilize contrastive learning to learn view-consistent information and distinguish outliers by the degree of consistency.
Experimental results on four benchmark datasets demonstrate that our proposed approach could outperform state-of-the-art competitors.
arXiv Detail & Related papers (2024-08-02T14:34:27Z) - Dual-View Data Hallucination with Semantic Relation Guidance for Few-Shot Image Recognition [49.26065739704278]
We propose a framework that exploits semantic relations to guide dual-view data hallucination for few-shot image recognition.
An instance-view data hallucination module hallucinates each sample of a novel class to generate new data.
A prototype-view data hallucination module exploits semantic-aware measure to estimate the prototype of a novel class.
arXiv Detail & Related papers (2024-01-13T12:32:29Z) - VDC: Versatile Data Cleanser based on Visual-Linguistic Inconsistency by Multimodal Large Language Models [46.72546879204724]
In the real-world, datasets may contain dirty samples, such as poisoned samples from backdoor attack, noisy labels in crowdsourcing, and even hybrids of them.
Existing detectors only focus on detecting poisoned samples or noisy labels, that are often prone to weak generalization when dealing with dirty samples from other domains.
We propose versatile data cleanser (VDC) leveraging the surpassing capabilities of multimodal large language models (MLLM) in cross-modal alignment and reasoning.
arXiv Detail & Related papers (2023-09-28T07:37:18Z) - Learning Noise-Robust Joint Representation for Multimodal Emotion Recognition under Incomplete Data Scenarios [23.43319138048058]
Multimodal emotion recognition (MER) in practical scenarios is significantly challenged by the presence of missing or incomplete data.
Traditional methods have often involved discarding data or substituting data segments with zero vectors to approximate these incompletenesses.
We introduce a novel noise-robust MER model that effectively learns robust multimodal joint representations from noisy data.
arXiv Detail & Related papers (2023-09-21T10:49:02Z) - Investigating and Mitigating the Side Effects of Noisy Views for Self-Supervised Clustering Algorithms in Practical Multi-View Scenarios [35.32285779434823]
Multi-view clustering (MVC) aims at exploring category structures among multi-view data in self-supervised manners.
noisy views might seriously degenerate when the views are noisy in practical multi-view scenarios.
We propose a theoretically grounded deep MVC method (namely MVCAN) to address this issue.
arXiv Detail & Related papers (2023-03-30T09:22:17Z) - Credible Remote Sensing Scene Classification Using Evidential Fusion on
Aerial-Ground Dual-view Images [6.817740582240199]
Multi-view (multi-source, multi-modal, multi-perspective, etc.) data are being used more frequently in remote sensing tasks.
The issue of data quality becomes more apparent, limiting the potential benefits of multi-view data.
Deep learning is introduced to the task of aerial-ground dual-view remote sensing scene classification.
arXiv Detail & Related papers (2023-01-02T12:27:55Z) - Reliable Shot Identification for Complex Event Detection via
Visual-Semantic Embedding [72.9370352430965]
We propose a visual-semantic guided loss method for event detection in videos.
Motivated by curriculum learning, we introduce a negative elastic regularization term to start training the classifier with instances of high reliability.
An alternative optimization algorithm is developed to solve the proposed challenging non-net regularization problem.
arXiv Detail & Related papers (2021-10-12T11:46:56Z) - Deep Partial Multi-View Learning [94.39367390062831]
We propose a novel framework termed Cross Partial Multi-View Networks (CPM-Nets)
We fifirst provide a formal defifinition of completeness and versatility for multi-view representation.
We then theoretically prove the versatility of the learned latent representations.
arXiv Detail & Related papers (2020-11-12T02:29:29Z) - Fully Unsupervised Diversity Denoising with Convolutional Variational
Autoencoders [81.30960319178725]
We propose DivNoising, a denoising approach based on fully convolutional variational autoencoders (VAEs)
First we introduce a principled way of formulating the unsupervised denoising problem within the VAE framework by explicitly incorporating imaging noise models into the decoder.
We show that such a noise model can either be measured, bootstrapped from noisy data, or co-learned during training.
arXiv Detail & Related papers (2020-06-10T21:28:13Z) - Noise Estimation Using Density Estimation for Self-Supervised Multimodal
Learning [10.151012770913624]
We show that noise estimation for multimodal data can be reduced to a multimodal density estimation task.
We demonstrate how our noise estimation can be broadly integrated and achieves comparable results to state-of-the-art performance.
arXiv Detail & Related papers (2020-03-06T13:25:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.