Central Similarity Multi-View Hashing for Multimedia Retrieval
- URL: http://arxiv.org/abs/2308.13774v1
- Date: Sat, 26 Aug 2023 05:43:29 GMT
- Title: Central Similarity Multi-View Hashing for Multimedia Retrieval
- Authors: Jian Zhu, Wen Cheng, Yu Cui, Chang Tang, Yuyang Dai, Yong Li, Lingfang
Zeng
- Abstract summary: We present a novel Central Similarity Multi-View Hashing (CSMVH) method to address the mentioned problems.
On the MS COCO and NUS-WIDE, the proposed CSMVH performs better than the state-of-the-art methods by a large margin.
- Score: 14.766486538338498
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hash representation learning of multi-view heterogeneous data is the key to
improving the accuracy of multimedia retrieval. However, existing methods
utilize local similarity and fall short of deeply fusing the multi-view
features, resulting in poor retrieval accuracy. Current methods only use local
similarity to train their model. These methods ignore global similarity.
Furthermore, most recent works fuse the multi-view features via a weighted sum
or concatenation. We contend that these fusion methods are insufficient for
capturing the interaction between various views. We present a novel Central
Similarity Multi-View Hashing (CSMVH) method to address the mentioned problems.
Central similarity learning is used for solving the local similarity problem,
which can utilize the global similarity between the hash center and samples. We
present copious empirical data demonstrating the superiority of gate-based
fusion over conventional approaches. On the MS COCO and NUS-WIDE, the proposed
CSMVH performs better than the state-of-the-art methods by a large margin (up
to 11.41% mean Average Precision (mAP) improvement).
Related papers
- DealMVC: Dual Contrastive Calibration for Multi-view Clustering [78.54355167448614]
We propose a novel Dual contrastive calibration network for Multi-View Clustering (DealMVC)
We first design a fusion mechanism to obtain a global cross-view feature. Then, a global contrastive calibration loss is proposed by aligning the view feature similarity graph and the high-confidence pseudo-label graph.
During the training procedure, the interacted cross-view feature is jointly optimized at both local and global levels.
arXiv Detail & Related papers (2023-08-17T14:14:28Z) - A Comparative Assessment of Multi-view fusion learning for Crop
Classification [3.883984493622102]
This work assesses different fusion strategies for crop classification in the CropHarvest dataset.
We present a comparison of multi-view fusion methods for three different datasets and show that, depending on the test region, different methods obtain the best performance.
arXiv Detail & Related papers (2023-08-10T08:03:58Z) - Deep Metric Multi-View Hashing for Multimedia Retrieval [3.539519688102545]
We propose a novel deep metric multi-view hashing (DMMVH) method to address the mentioned problems.
On the MIR-Flickr25K, MS COCO, and NUS-WIDE, our method outperforms the current state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2023-04-13T09:25:35Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Decoupling Predictions in Distributed Learning for Multi-Center Left
Atrial MRI Segmentation [20.20518948616193]
We propose a new framework of distributed learning that bridges the gap between two groups, and improves the performance for both generic and local data.
Results on multi-center left atrial (LA) MRI segmentation showed that our method demonstrated superior performance over existing methods on both generic and local data.
arXiv Detail & Related papers (2022-06-10T08:35:42Z) - Instance-weighted Central Similarity for Multi-label Image Retrieval [66.23348499938278]
We propose Instance-weighted Central Similarity (ICS) to automatically learn the center weight corresponding to a hash code.
Our method achieves the state-of-the-art performance on the image retrieval benchmarks, and especially improves the mAP by 1.6%-6.4% on the MS COCO dataset.
arXiv Detail & Related papers (2021-08-11T15:18:18Z) - Deep Multimodal Fusion by Channel Exchanging [87.40768169300898]
This paper proposes a parameter-free multimodal fusion framework that dynamically exchanges channels between sub-networks of different modalities.
The validity of such exchanging process is also guaranteed by sharing convolutional filters yet keeping separate BN layers across modalities, which, as an add-on benefit, allows our multimodal architecture to be almost as compact as a unimodal network.
arXiv Detail & Related papers (2020-11-10T09:53:20Z) - CIMON: Towards High-quality Hash Codes [63.37321228830102]
We propose a new method named textbfComprehensive stextbfImilarity textbfMining and ctextbfOnsistency leartextbfNing (CIMON)
First, we use global refinement and similarity statistical distribution to obtain reliable and smooth guidance. Second, both semantic and contrastive consistency learning are introduced to derive both disturb-invariant and discriminative hash codes.
arXiv Detail & Related papers (2020-10-15T14:47:14Z) - Multi-Center Federated Learning [62.57229809407692]
This paper proposes a novel multi-center aggregation mechanism for federated learning.
It learns multiple global models from the non-IID user data and simultaneously derives the optimal matching between users and centers.
Our experimental results on benchmark datasets show that our method outperforms several popular federated learning methods.
arXiv Detail & Related papers (2020-05-03T09:14:31Z) - Deep Multi-View Enhancement Hashing for Image Retrieval [40.974719473643724]
This paper proposes a supervised multi-view hash model which can enhance the multi-view information through neural networks.
The proposed method is systematically evaluated on the CIFAR-10, NUS-WIDE and MS-COCO datasets.
arXiv Detail & Related papers (2020-02-01T08:32:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.